var/home/core/zuul-output/0000755000175000017500000000000015116572731014535 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116576245015505 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000001676400015116576236017715 0ustar rootrootDec 11 16:52:42 crc systemd[1]: Starting Kubernetes Kubelet... Dec 11 16:52:42 crc kubenswrapper[5109]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 16:52:42 crc kubenswrapper[5109]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 11 16:52:42 crc kubenswrapper[5109]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 16:52:42 crc kubenswrapper[5109]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 16:52:42 crc kubenswrapper[5109]: Flag --pod-infra-container-image has been deprecated, will be removed in 1.35. Image garbage collector will get sandbox image information from CRI. Dec 11 16:52:42 crc kubenswrapper[5109]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.640293 5109 server.go:212] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645275 5109 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645309 5109 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645318 5109 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645329 5109 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645339 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645348 5109 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645356 5109 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645365 5109 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645374 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645383 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645390 5109 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645398 5109 feature_gate.go:328] unrecognized feature gate: PinnedImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645407 5109 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645415 5109 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645422 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645431 5109 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645438 5109 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645446 5109 feature_gate.go:328] unrecognized feature gate: DualReplica Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645454 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645462 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645470 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645478 5109 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645487 5109 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645496 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645505 5109 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645515 5109 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645525 5109 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645535 5109 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645545 5109 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645555 5109 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645565 5109 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645575 5109 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645585 5109 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645595 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645606 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645616 5109 feature_gate.go:328] unrecognized feature gate: Example2 Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645627 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645638 5109 feature_gate.go:328] unrecognized feature gate: NewOLM Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645648 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645659 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645667 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645677 5109 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645687 5109 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645698 5109 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645709 5109 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645718 5109 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645728 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645771 5109 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645780 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645788 5109 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645800 5109 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645812 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645822 5109 feature_gate.go:328] unrecognized feature gate: SignatureStores Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645832 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645843 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645857 5109 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645873 5109 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645883 5109 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645893 5109 feature_gate.go:328] unrecognized feature gate: OVNObservability Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645904 5109 feature_gate.go:328] unrecognized feature gate: Example Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645911 5109 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645920 5109 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645930 5109 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645940 5109 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645950 5109 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645960 5109 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645969 5109 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645979 5109 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645989 5109 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.645999 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646009 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646018 5109 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646030 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646045 5109 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646057 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646068 5109 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646083 5109 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646093 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646102 5109 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646111 5109 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646121 5109 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646131 5109 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646141 5109 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646151 5109 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646161 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.646172 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.647957 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.647978 5109 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.647988 5109 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.647996 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648005 5109 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648014 5109 feature_gate.go:328] unrecognized feature gate: OVNObservability Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648023 5109 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648070 5109 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648081 5109 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648091 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648101 5109 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648109 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648118 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648126 5109 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648133 5109 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648142 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648150 5109 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648159 5109 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648168 5109 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648177 5109 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648186 5109 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648193 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648201 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648209 5109 feature_gate.go:328] unrecognized feature gate: PinnedImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648217 5109 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648225 5109 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648233 5109 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648241 5109 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648249 5109 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648257 5109 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648266 5109 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648274 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648282 5109 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648289 5109 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648298 5109 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648305 5109 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648314 5109 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648321 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648329 5109 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648336 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648344 5109 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648352 5109 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648359 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648367 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648376 5109 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648384 5109 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648392 5109 feature_gate.go:328] unrecognized feature gate: Example2 Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648400 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648408 5109 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648419 5109 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648428 5109 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648438 5109 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648446 5109 feature_gate.go:328] unrecognized feature gate: SignatureStores Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648454 5109 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648462 5109 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648470 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648477 5109 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648486 5109 feature_gate.go:328] unrecognized feature gate: DualReplica Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648494 5109 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648502 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648510 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648517 5109 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648526 5109 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648534 5109 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648541 5109 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648549 5109 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648557 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648565 5109 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648572 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648580 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648587 5109 feature_gate.go:328] unrecognized feature gate: NewOLM Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648597 5109 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648605 5109 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648613 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648621 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648628 5109 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648636 5109 feature_gate.go:328] unrecognized feature gate: Example Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648643 5109 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648651 5109 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648659 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648666 5109 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648674 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648681 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648689 5109 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648698 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.648705 5109 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648910 5109 flags.go:64] FLAG: --address="0.0.0.0" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648926 5109 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648941 5109 flags.go:64] FLAG: --anonymous-auth="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648952 5109 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648964 5109 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648974 5109 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648987 5109 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.648999 5109 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649009 5109 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649018 5109 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649027 5109 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649036 5109 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649046 5109 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649056 5109 flags.go:64] FLAG: --cgroup-root="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649064 5109 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649073 5109 flags.go:64] FLAG: --client-ca-file="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649081 5109 flags.go:64] FLAG: --cloud-config="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649089 5109 flags.go:64] FLAG: --cloud-provider="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649098 5109 flags.go:64] FLAG: --cluster-dns="[]" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649111 5109 flags.go:64] FLAG: --cluster-domain="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649119 5109 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649128 5109 flags.go:64] FLAG: --config-dir="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649136 5109 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649145 5109 flags.go:64] FLAG: --container-log-max-files="5" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649156 5109 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649165 5109 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649173 5109 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649182 5109 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649191 5109 flags.go:64] FLAG: --contention-profiling="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649200 5109 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649208 5109 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649220 5109 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649229 5109 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649240 5109 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649249 5109 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649259 5109 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649267 5109 flags.go:64] FLAG: --enable-load-reader="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649276 5109 flags.go:64] FLAG: --enable-server="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649285 5109 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649296 5109 flags.go:64] FLAG: --event-burst="100" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649305 5109 flags.go:64] FLAG: --event-qps="50" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649314 5109 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649322 5109 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649331 5109 flags.go:64] FLAG: --eviction-hard="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649350 5109 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649359 5109 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649368 5109 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649377 5109 flags.go:64] FLAG: --eviction-soft="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649385 5109 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649394 5109 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649403 5109 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649411 5109 flags.go:64] FLAG: --experimental-mounter-path="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649419 5109 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649428 5109 flags.go:64] FLAG: --fail-swap-on="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649436 5109 flags.go:64] FLAG: --feature-gates="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649447 5109 flags.go:64] FLAG: --file-check-frequency="20s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649456 5109 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649465 5109 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649474 5109 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649482 5109 flags.go:64] FLAG: --healthz-port="10248" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649491 5109 flags.go:64] FLAG: --help="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649500 5109 flags.go:64] FLAG: --hostname-override="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649508 5109 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649517 5109 flags.go:64] FLAG: --http-check-frequency="20s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649533 5109 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649542 5109 flags.go:64] FLAG: --image-credential-provider-config="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649550 5109 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649559 5109 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649568 5109 flags.go:64] FLAG: --image-service-endpoint="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649576 5109 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649588 5109 flags.go:64] FLAG: --kube-api-burst="100" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649596 5109 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649605 5109 flags.go:64] FLAG: --kube-api-qps="50" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649614 5109 flags.go:64] FLAG: --kube-reserved="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649623 5109 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649633 5109 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649646 5109 flags.go:64] FLAG: --kubelet-cgroups="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649655 5109 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649663 5109 flags.go:64] FLAG: --lock-file="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649672 5109 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649681 5109 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649690 5109 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649703 5109 flags.go:64] FLAG: --log-json-split-stream="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649711 5109 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649720 5109 flags.go:64] FLAG: --log-text-split-stream="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649729 5109 flags.go:64] FLAG: --logging-format="text" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649765 5109 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649775 5109 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649784 5109 flags.go:64] FLAG: --manifest-url="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649792 5109 flags.go:64] FLAG: --manifest-url-header="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649804 5109 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649814 5109 flags.go:64] FLAG: --max-open-files="1000000" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649825 5109 flags.go:64] FLAG: --max-pods="110" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649834 5109 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649843 5109 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649851 5109 flags.go:64] FLAG: --memory-manager-policy="None" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649860 5109 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649871 5109 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649880 5109 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649888 5109 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhel" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649919 5109 flags.go:64] FLAG: --node-status-max-images="50" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649930 5109 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649942 5109 flags.go:64] FLAG: --oom-score-adj="-999" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649953 5109 flags.go:64] FLAG: --pod-cidr="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649963 5109 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:cc2b30e70040205c2536d01ae5c850be1ed2d775cf13249e50328e5085777977" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649982 5109 flags.go:64] FLAG: --pod-manifest-path="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.649994 5109 flags.go:64] FLAG: --pod-max-pids="-1" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650005 5109 flags.go:64] FLAG: --pods-per-core="0" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650016 5109 flags.go:64] FLAG: --port="10250" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650033 5109 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650044 5109 flags.go:64] FLAG: --provider-id="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650056 5109 flags.go:64] FLAG: --qos-reserved="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650068 5109 flags.go:64] FLAG: --read-only-port="10255" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650079 5109 flags.go:64] FLAG: --register-node="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650089 5109 flags.go:64] FLAG: --register-schedulable="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650100 5109 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650118 5109 flags.go:64] FLAG: --registry-burst="10" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650130 5109 flags.go:64] FLAG: --registry-qps="5" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650140 5109 flags.go:64] FLAG: --reserved-cpus="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650151 5109 flags.go:64] FLAG: --reserved-memory="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650162 5109 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650171 5109 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650180 5109 flags.go:64] FLAG: --rotate-certificates="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650193 5109 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650202 5109 flags.go:64] FLAG: --runonce="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650210 5109 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650219 5109 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650228 5109 flags.go:64] FLAG: --seccomp-default="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650237 5109 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650245 5109 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650256 5109 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650265 5109 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650274 5109 flags.go:64] FLAG: --storage-driver-password="root" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650282 5109 flags.go:64] FLAG: --storage-driver-secure="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650290 5109 flags.go:64] FLAG: --storage-driver-table="stats" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650299 5109 flags.go:64] FLAG: --storage-driver-user="root" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650307 5109 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650316 5109 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650325 5109 flags.go:64] FLAG: --system-cgroups="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650333 5109 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650347 5109 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650358 5109 flags.go:64] FLAG: --tls-cert-file="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650367 5109 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650378 5109 flags.go:64] FLAG: --tls-min-version="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650387 5109 flags.go:64] FLAG: --tls-private-key-file="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650395 5109 flags.go:64] FLAG: --topology-manager-policy="none" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650403 5109 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650412 5109 flags.go:64] FLAG: --topology-manager-scope="container" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650421 5109 flags.go:64] FLAG: --v="2" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650432 5109 flags.go:64] FLAG: --version="false" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650443 5109 flags.go:64] FLAG: --vmodule="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650453 5109 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.650462 5109 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650668 5109 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650677 5109 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650689 5109 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650698 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650706 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650714 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650721 5109 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650729 5109 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650772 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650780 5109 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650789 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650797 5109 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650805 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650812 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650821 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650829 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650837 5109 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650844 5109 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650852 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650859 5109 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650870 5109 feature_gate.go:328] unrecognized feature gate: NewOLM Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650878 5109 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650886 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650893 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650901 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650909 5109 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650917 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650924 5109 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650932 5109 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650939 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650947 5109 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650955 5109 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650963 5109 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650971 5109 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650982 5109 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650989 5109 feature_gate.go:328] unrecognized feature gate: SignatureStores Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.650998 5109 feature_gate.go:328] unrecognized feature gate: Example2 Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651006 5109 feature_gate.go:328] unrecognized feature gate: DualReplica Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651014 5109 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651022 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651029 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651037 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651048 5109 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651067 5109 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651076 5109 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651085 5109 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651093 5109 feature_gate.go:328] unrecognized feature gate: PinnedImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651102 5109 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651110 5109 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651117 5109 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651125 5109 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651132 5109 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651143 5109 feature_gate.go:328] unrecognized feature gate: Example Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651151 5109 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651159 5109 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651167 5109 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651174 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651182 5109 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651190 5109 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651200 5109 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651210 5109 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651218 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651226 5109 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651234 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651242 5109 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651250 5109 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651261 5109 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651269 5109 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651277 5109 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651285 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651292 5109 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651300 5109 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651308 5109 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651316 5109 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651324 5109 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651332 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651340 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651376 5109 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651386 5109 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651395 5109 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651404 5109 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651413 5109 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651422 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651430 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651441 5109 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.651448 5109 feature_gate.go:328] unrecognized feature gate: OVNObservability Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.651688 5109 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.661424 5109 server.go:530] "Kubelet version" kubeletVersion="v1.33.5" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.661443 5109 server.go:532] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661501 5109 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661507 5109 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661513 5109 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661518 5109 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661522 5109 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661527 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661532 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661536 5109 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661541 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661545 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661549 5109 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661554 5109 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661558 5109 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661563 5109 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661567 5109 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661572 5109 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661576 5109 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661581 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661585 5109 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661591 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661595 5109 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661600 5109 feature_gate.go:328] unrecognized feature gate: OVNObservability Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661604 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661611 5109 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661617 5109 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661623 5109 feature_gate.go:328] unrecognized feature gate: NewOLM Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661628 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661632 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661638 5109 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661644 5109 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661649 5109 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661655 5109 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661660 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661665 5109 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661669 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661674 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661678 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661683 5109 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661687 5109 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661692 5109 feature_gate.go:328] unrecognized feature gate: Example2 Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661696 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661700 5109 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661705 5109 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661710 5109 feature_gate.go:328] unrecognized feature gate: PinnedImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661714 5109 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661718 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661723 5109 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661727 5109 feature_gate.go:328] unrecognized feature gate: DualReplica Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661731 5109 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661752 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661757 5109 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661762 5109 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661767 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661772 5109 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661776 5109 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661781 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661785 5109 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661790 5109 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661794 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661799 5109 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661803 5109 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661808 5109 feature_gate.go:328] unrecognized feature gate: SignatureStores Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661813 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661819 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661823 5109 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661828 5109 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661832 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661837 5109 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661841 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661846 5109 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661850 5109 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661854 5109 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661859 5109 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661864 5109 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661868 5109 feature_gate.go:328] unrecognized feature gate: Example Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661873 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661877 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661882 5109 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661886 5109 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661890 5109 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661895 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661899 5109 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661904 5109 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661908 5109 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661913 5109 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.661918 5109 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.661925 5109 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662069 5109 feature_gate.go:328] unrecognized feature gate: PinnedImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662077 5109 feature_gate.go:328] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662082 5109 feature_gate.go:328] unrecognized feature gate: AdditionalRoutingCapabilities Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662087 5109 feature_gate.go:328] unrecognized feature gate: ImageModeStatusReporting Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662092 5109 feature_gate.go:328] unrecognized feature gate: AzureMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662096 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662101 5109 feature_gate.go:328] unrecognized feature gate: Example Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662107 5109 feature_gate.go:328] unrecognized feature gate: MachineConfigNodes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662112 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpointsInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662117 5109 feature_gate.go:328] unrecognized feature gate: DualReplica Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662122 5109 feature_gate.go:328] unrecognized feature gate: VolumeGroupSnapshot Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662127 5109 feature_gate.go:328] unrecognized feature gate: VSphereHostVMGroupZonal Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662132 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662137 5109 feature_gate.go:328] unrecognized feature gate: ExternalSnapshotMetadata Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662141 5109 feature_gate.go:328] unrecognized feature gate: HighlyAvailableArbiter Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662146 5109 feature_gate.go:328] unrecognized feature gate: DyanmicServiceEndpointIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662150 5109 feature_gate.go:328] unrecognized feature gate: VSphereConfigurableMaxAllowedBlockVolumesPerNode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662155 5109 feature_gate.go:328] unrecognized feature gate: ImageStreamImportMode Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662160 5109 feature_gate.go:328] unrecognized feature gate: IrreconcilableMachineConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662164 5109 feature_gate.go:328] unrecognized feature gate: ClusterVersionOperatorConfiguration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662169 5109 feature_gate.go:328] unrecognized feature gate: AzureWorkloadIdentity Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662173 5109 feature_gate.go:328] unrecognized feature gate: NewOLMOwnSingleNamespace Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662178 5109 feature_gate.go:328] unrecognized feature gate: AzureDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662183 5109 feature_gate.go:328] unrecognized feature gate: NewOLM Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662187 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662194 5109 feature_gate.go:351] Setting GA feature gate ServiceAccountTokenNodeBinding=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662199 5109 feature_gate.go:328] unrecognized feature gate: GCPCustomAPIEndpoints Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662205 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerificationPKI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662209 5109 feature_gate.go:328] unrecognized feature gate: CPMSMachineNamePrefix Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662214 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfigAPI Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662218 5109 feature_gate.go:328] unrecognized feature gate: NetworkLiveMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662223 5109 feature_gate.go:328] unrecognized feature gate: GatewayAPIController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662228 5109 feature_gate.go:328] unrecognized feature gate: MixedCPUsAllocation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662233 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662237 5109 feature_gate.go:328] unrecognized feature gate: UpgradeStatus Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662242 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiDisk Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662246 5109 feature_gate.go:328] unrecognized feature gate: BootcNodeManagement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662251 5109 feature_gate.go:328] unrecognized feature gate: PreconfiguredUDNAddresses Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662255 5109 feature_gate.go:328] unrecognized feature gate: AWSServiceLBNetworkSecurityGroup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662260 5109 feature_gate.go:328] unrecognized feature gate: VSphereMixedNodeEnv Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662265 5109 feature_gate.go:328] unrecognized feature gate: MetricsCollectionProfiles Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662269 5109 feature_gate.go:328] unrecognized feature gate: BuildCSIVolumes Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662274 5109 feature_gate.go:328] unrecognized feature gate: BootImageSkewEnforcement Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662279 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662283 5109 feature_gate.go:328] unrecognized feature gate: NutanixMultiSubnets Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662288 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662292 5109 feature_gate.go:328] unrecognized feature gate: NewOLMPreflightPermissionChecks Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662297 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImages Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662302 5109 feature_gate.go:328] unrecognized feature gate: DNSNameResolver Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662307 5109 feature_gate.go:328] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662312 5109 feature_gate.go:328] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662317 5109 feature_gate.go:328] unrecognized feature gate: InsightsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662322 5109 feature_gate.go:328] unrecognized feature gate: RouteAdvertisements Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662327 5109 feature_gate.go:328] unrecognized feature gate: EtcdBackendQuota Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662332 5109 feature_gate.go:328] unrecognized feature gate: AutomatedEtcdBackup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662337 5109 feature_gate.go:328] unrecognized feature gate: AzureClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662343 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIMigration Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662348 5109 feature_gate.go:328] unrecognized feature gate: MultiArchInstallAzure Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662353 5109 feature_gate.go:328] unrecognized feature gate: NetworkSegmentation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662358 5109 feature_gate.go:328] unrecognized feature gate: ClusterMonitoringConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662362 5109 feature_gate.go:328] unrecognized feature gate: SigstoreImageVerification Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662367 5109 feature_gate.go:328] unrecognized feature gate: OVNObservability Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662372 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDC Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662376 5109 feature_gate.go:328] unrecognized feature gate: GCPClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662382 5109 feature_gate.go:328] unrecognized feature gate: NoRegistryClusterOperations Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662386 5109 feature_gate.go:328] unrecognized feature gate: NewOLMWebhookProviderOpenshiftServiceCA Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662391 5109 feature_gate.go:328] unrecognized feature gate: AdminNetworkPolicy Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662396 5109 feature_gate.go:328] unrecognized feature gate: AWSDedicatedHosts Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662400 5109 feature_gate.go:328] unrecognized feature gate: SetEIPForNLBIngressController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662405 5109 feature_gate.go:328] unrecognized feature gate: MultiDiskSetup Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662410 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNSInstall Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662415 5109 feature_gate.go:328] unrecognized feature gate: ShortCertRotation Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662420 5109 feature_gate.go:328] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662424 5109 feature_gate.go:328] unrecognized feature gate: ManagedBootImagesvSphere Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662429 5109 feature_gate.go:328] unrecognized feature gate: SignatureStores Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662434 5109 feature_gate.go:328] unrecognized feature gate: AlibabaPlatform Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662438 5109 feature_gate.go:328] unrecognized feature gate: InsightsOnDemandDataGather Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662443 5109 feature_gate.go:328] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662447 5109 feature_gate.go:328] unrecognized feature gate: NetworkDiagnosticsConfig Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662452 5109 feature_gate.go:328] unrecognized feature gate: KMSEncryptionProvider Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662458 5109 feature_gate.go:349] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662464 5109 feature_gate.go:328] unrecognized feature gate: Example2 Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662469 5109 feature_gate.go:328] unrecognized feature gate: ExternalOIDCWithUIDAndExtraClaimMappings Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662473 5109 feature_gate.go:328] unrecognized feature gate: AWSClusterHostedDNS Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662478 5109 feature_gate.go:328] unrecognized feature gate: NewOLMCatalogdAPIV1Metas Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.662482 5109 feature_gate.go:328] unrecognized feature gate: VSphereMultiNetworks Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.662490 5109 feature_gate.go:384] feature gates: {map[DynamicResourceAllocation:false EventedPLEG:false ImageVolume:true KMSv1:true MaxUnavailableStatefulSet:false MinimumKubeletVersion:false MutatingAdmissionPolicy:false NodeSwap:false ProcMountType:true RouteExternalCertificate:true SELinuxMount:false ServiceAccountTokenNodeBinding:true StoragePerformantSecurityPolicy:true TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:true UserNamespacesSupport:true VolumeAttributesClass:false]} Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.662856 5109 server.go:962] "Client rotation is on, will bootstrap in background" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.667471 5109 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2025-12-03 08:27:53 +0000 UTC" logger="UnhandledError" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.672731 5109 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.672909 5109 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-client-current.pem" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.673706 5109 server.go:1019] "Starting client certificate rotation" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.673905 5109 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kube-apiserver-client-kubelet" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.673996 5109 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.686792 5109 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.690099 5109 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.693844 5109 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.710056 5109 log.go:25] "Validated CRI v1 runtime API" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.743854 5109 log.go:25] "Validated CRI v1 image API" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.747585 5109 server.go:1452] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.750782 5109 fs.go:135] Filesystem UUIDs: map[19e76f87-96b8-4794-9744-0b33dca22d5b:/dev/vda3 2025-12-11-16-46-33-00:/dev/sr0 5eb7c122-420e-4494-80ec-41664070d7b6:/dev/vda4 7B77-95E7:/dev/vda2] Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.750846 5109 fs.go:136] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:45 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:31 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:46 fsType:tmpfs blockSize:0} composefs_0-33:{mountpoint:/ major:0 minor:33 fsType:overlay blockSize:0}] Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.779379 5109 manager.go:217] Machine: {Timestamp:2025-12-11 16:52:42.777348674 +0000 UTC m=+0.457040190 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33649930240 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:80bc4fba336e4ca1bc9d28a8be52a356 SystemUUID:789f6dfb-7431-471d-a595-070518eb3889 BootID:2bb2c55e-c8c8-470b-a3e3-d0e526c54090 Filesystems:[{Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:45 Capacity:3364990976 Type:vfs Inodes:821531 HasInodes:true} {Device:composefs_0-33 DeviceMajor:0 DeviceMinor:33 Capacity:6545408 Type:vfs Inodes:18446744073709551615 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:31 Capacity:16824967168 Type:vfs Inodes:1048576 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:46 Capacity:1073741824 Type:vfs Inodes:4107657 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16824963072 Type:vfs Inodes:4107657 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6729986048 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:b8:a1:c3 Speed:0 Mtu:1500} {Name:br-int MacAddress:b2:a9:9f:57:07:84 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:b8:a1:c3 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:24:ed:96 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:56:16:ca Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:50:89:16 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:54:92:84 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:12:d7:6c:ce:fc:9a Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f6:44:68:51:2d:95 Speed:0 Mtu:1500} {Name:tap0 MacAddress:5a:94:ef:e4:0c:ee Speed:10 Mtu:1500}] Topology:[{Id:0 Memory:33649930240 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.779913 5109 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.780153 5109 manager.go:233] Version: {KernelVersion:5.14.0-570.57.1.el9_6.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 9.6.20251021-0 (Plow) DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.782287 5109 container_manager_linux.go:270] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.782345 5109 container_manager_linux.go:275] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"MemoryManagerPolicy":"None","MemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.782593 5109 topology_manager.go:138] "Creating topology manager with none policy" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.782605 5109 container_manager_linux.go:306] "Creating device plugin manager" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.782632 5109 manager.go:141] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.783789 5109 server.go:72] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.784311 5109 state_mem.go:36] "Initialized new in-memory state store" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.784512 5109 server.go:1267] "Using root directory" path="/var/lib/kubelet" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.785874 5109 kubelet.go:491] "Attempting to sync node with API server" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.785906 5109 kubelet.go:386] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.785924 5109 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.785943 5109 kubelet.go:397] "Adding apiserver pod source" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.785962 5109 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.791275 5109 state_checkpoint.go:81] "State checkpoint: restored pod resource state from checkpoint" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.791381 5109 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.795675 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.796389 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.798465 5109 state_checkpoint.go:81] "State checkpoint: restored pod resource state from checkpoint" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.798520 5109 state_mem.go:40] "Initialized new in-memory state store for pod resource information tracking" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.803048 5109 kuberuntime_manager.go:279] "Container runtime initialized" containerRuntime="cri-o" version="1.33.5-3.rhaos4.20.gitd0ea985.el9" apiVersion="v1" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.803281 5109 certificate_store.go:147] "Loading cert/key pair from a file" filePath="/var/lib/kubelet/pki/kubelet-server-current.pem" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804058 5109 kubelet.go:953] "Not starting ClusterTrustBundle informer because we are in static kubelet mode or the ClusterTrustBundleProjection featuregate is disabled" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804510 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804540 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804550 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804594 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804604 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804613 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804624 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804633 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804643 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804659 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804673 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.804837 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.805076 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.805087 5109 plugins.go:616] "Loaded volume plugin" pluginName="kubernetes.io/image" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.807374 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.818133 5109 watchdog_linux.go:99] "Systemd watchdog is not enabled" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.818237 5109 server.go:1295] "Started kubelet" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.818530 5109 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.818725 5109 server_v1.go:47] "podresources" method="list" useActivePods=true Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.818837 5109 server.go:180] "Starting to listen" address="0.0.0.0" port=10250 Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.820089 5109 server.go:255] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 11 16:52:42 crc systemd[1]: Started Kubernetes Kubelet. Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.821096 5109 server.go:317] "Adding debug handlers to kubelet server" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.821403 5109 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1880376574b07590 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.818180496 +0000 UTC m=+0.497871972,LastTimestamp:2025-12-11 16:52:42.818180496 +0000 UTC m=+0.497871972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.822810 5109 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.822818 5109 certificate_manager.go:422] "Certificate rotation is enabled" logger="kubernetes.io/kubelet-serving" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.824250 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.824498 5109 volume_manager.go:295] "The desired_state_of_world populator starts" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.824572 5109 volume_manager.go:297] "Starting Kubelet Volume Manager" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.824834 5109 desired_state_of_world_populator.go:150] "Desired state populator starts to run" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.825795 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.825856 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.834081 5109 factory.go:55] Registering systemd factory Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.834133 5109 factory.go:223] Registration of the systemd container factory successfully Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.835460 5109 factory.go:153] Registering CRI-O factory Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.835489 5109 factory.go:223] Registration of the crio container factory successfully Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.835578 5109 factory.go:221] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.835796 5109 factory.go:103] Registering Raw factory Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.835819 5109 manager.go:1196] Started watching for new ooms in manager Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.836789 5109 manager.go:319] Starting recovery of all containers Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870780 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870834 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f863fff9-286a-45fa-b8f0-8a86994b8440" volumeName="kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870872 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870885 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870898 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870909 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870920 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870931 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870945 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870958 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870969 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870982 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.870993 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="17b87002-b798-480a-8e17-83053d698239" volumeName="kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871003 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5ebfebf6-3ecd-458e-943f-bb25b52e2718" volumeName="kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871018 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af41de71-79cf-4590-bbe9-9e8b848862cb" volumeName="kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871028 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871041 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871080 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871091 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871104 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871115 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871126 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871137 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871149 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871166 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871178 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871188 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871201 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" volumeName="kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871273 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871288 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871303 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871316 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871367 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871385 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871399 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a14caf222afb62aaabdc47808b6f944" volumeName="kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871413 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ce090a97-9ab6-4c40-a719-64ff2acd9778" volumeName="kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871426 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871439 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871454 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871468 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871486 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871500 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871515 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871613 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871629 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871644 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871660 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" volumeName="kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871699 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871713 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.871727 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.872869 5109 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b1264ac67579ad07e7e9003054d44fe40dd55285a4b2f7dc74e48be1aee0868a/globalmount" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.872901 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.872918 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.872933 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.872946 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.872961 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.872976 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873004 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873018 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873064 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873108 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873124 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873138 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873117 5109 manager.go:324] Recovery completed Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873160 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873858 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873916 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.873966 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874019 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874078 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874103 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874129 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874156 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874178 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f0bc7fcb0822a2c13eb2d22cd8c0641" volumeName="kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874197 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874217 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874237 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874260 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874278 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874296 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874315 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874333 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874353 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874372 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874390 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874409 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874429 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3a14caf222afb62aaabdc47808b6f944" volumeName="kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874464 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874483 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874503 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874523 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874540 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874560 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874578 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874598 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874618 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874640 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874659 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874679 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7e2c886-118e-43bb-bef1-c78134de392b" volumeName="kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874700 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874728 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874786 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874806 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7599e0b6-bddf-4def-b7f2-0b32206e8651" volumeName="kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874827 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a208c9c2-333b-4b4a-be0d-bc32ec38a821" volumeName="kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874846 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" volumeName="kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874865 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874884 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874902 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874925 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="94a6e063-3d1a-4d44-875d-185291448c31" volumeName="kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874946 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874966 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e093be35-bb62-4843-b2e8-094545761610" volumeName="kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.874996 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875015 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875035 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875084 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875116 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875133 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d7e8f42f-dc0e-424b-bb56-5ec849834888" volumeName="kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875152 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="34177974-8d82-49d2-a763-391d0df3bbd8" volumeName="kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875171 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="34177974-8d82-49d2-a763-391d0df3bbd8" volumeName="kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875229 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875249 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875267 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875286 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875304 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="584e1f4a-8205-47d7-8efb-3afc6017c4c9" volumeName="kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875321 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875339 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875367 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875388 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875408 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9f71a554-e414-4bc3-96d2-674060397afe" volumeName="kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875426 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875444 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875465 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875482 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31fa8943-81cc-4750-a0b7-0fa9ab5af883" volumeName="kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875509 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875536 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875556 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875575 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7e2c886-118e-43bb-bef1-c78134de392b" volumeName="kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875594 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875613 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875632 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875649 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875670 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875689 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875719 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875760 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875780 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" volumeName="kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875798 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875817 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875835 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875853 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875875 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875894 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875913 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a208c9c2-333b-4b4a-be0d-bc32ec38a821" volumeName="kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875932 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" volumeName="kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875951 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875969 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="42a11a02-47e1-488f-b270-2679d3298b0e" volumeName="kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.875986 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876004 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876023 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876046 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876070 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="42a11a02-47e1-488f-b270-2679d3298b0e" volumeName="kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876095 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876119 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876141 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876173 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876191 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876208 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876228 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876248 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876268 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876292 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876316 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876341 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876370 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" volumeName="kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876396 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876421 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876446 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876470 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876494 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876522 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" volumeName="kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876546 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876572 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876589 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876608 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876631 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f559dfa3-3917-43a2-97f6-61ddfda10e93" volumeName="kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876655 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876681 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876705 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876732 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876789 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a555ff2e-0be6-46d5-897d-863bb92ae2b3" volumeName="kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876808 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a7a88189-c967-4640-879e-27665747f20c" volumeName="kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876829 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" volumeName="kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876853 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20c5c5b4bed930554494851fe3cb2b2a" volumeName="kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876898 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="567683bd-0efc-4f21-b076-e28559628404" volumeName="kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876917 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="593a3561-7760-45c5-8f91-5aaef7475d0f" volumeName="kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876935 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" volumeName="kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876954 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" volumeName="kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876973 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c5f2bfad-70f6-4185-a3d9-81ce12720767" volumeName="kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.876991 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f65c0ac1-8bca-454d-a2e6-e35cb418beac" volumeName="kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877009 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01080b46-74f1-4191-8755-5152a57b3b25" volumeName="kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877038 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877058 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="149b3c48-e17c-4a66-a835-d86dabf6ff13" volumeName="kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877076 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="301e1965-1754-483d-b6cc-bfae7038bbca" volumeName="kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877093 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877111 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877133 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877152 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6077b63e-53a2-4f96-9d56-1ce0324e4913" volumeName="kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877171 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877190 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7df94c10-441d-4386-93a6-6730fb7bcde0" volumeName="kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877207 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d19cb085-0c5b-4810-b654-ce7923221d90" volumeName="kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877226 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5ebfebf6-3ecd-458e-943f-bb25b52e2718" volumeName="kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877245 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" volumeName="kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877265 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ee8fbd3-1f81-4666-96da-5afc70819f1a" volumeName="kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877285 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ee8fbd3-1f81-4666-96da-5afc70819f1a" volumeName="kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877312 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" volumeName="kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877336 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d565531a-ff86-4608-9d19-767de01ac31b" volumeName="kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877361 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877386 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" volumeName="kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877407 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b638b8f4bb0070e40528db779baf6a2" volumeName="kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877423 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="2325ffef-9d5b-447f-b00e-3efc429acefe" volumeName="kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877442 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0effdbcf-dd7d-404d-9d48-77536d665a5d" volumeName="kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877461 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="18f80adb-c1c3-49ba-8ee4-932c851d3897" volumeName="kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877480 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" volumeName="kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877500 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" volumeName="kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877518 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="736c54fe-349c-4bb9-870a-d1c1d1c03831" volumeName="kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877535 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="92dfbade-90b6-4169-8c07-72cff7f2c82b" volumeName="kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877610 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b605f283-6f2e-42da-a838-54421690f7d0" volumeName="kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877627 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" volumeName="kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877643 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" volumeName="kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877663 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="869851b9-7ffb-4af0-b166-1d8aa40a5f80" volumeName="kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877682 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="af33e427-6803-48c2-a76a-dd9deb7cbf9a" volumeName="kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877698 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c491984c-7d4b-44aa-8c1e-d7974424fa47" volumeName="kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877715 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fc4541ce-7789-4670-bc75-5c2868e52ce0" volumeName="kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877786 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="16bdd140-dce1-464c-ab47-dd5798d1d256" volumeName="kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877815 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7afa918d-be67-40a6-803c-d3b0ae99d815" volumeName="kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877833 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="81e39f7b-62e4-4fc9-992a-6535ce127a02" volumeName="kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877850 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9e9b5059-1b3e-4067-a63d-2952cbe863af" volumeName="kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877869 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cc85e424-18b2-4924-920b-bd291a8c4b01" volumeName="kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877886 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" volumeName="kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877907 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6edfcf45-925b-4eff-b940-95b6fc0b85d4" volumeName="kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877925 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" volumeName="kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877943 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" volumeName="kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877960 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09cfa50b-4138-4585-a53e-64dd3ab73335" volumeName="kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877976 5109 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b4750666-1362-4001-abd0-6f89964cc621" volumeName="kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config" seLinuxMountContext="" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.877994 5109 reconstruct.go:97] "Volume reconstruction finished" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.878009 5109 reconciler.go:26] "Reconciler: start to sync state" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.895910 5109 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv4" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.898859 5109 kubelet_network_linux.go:49] "Initialized iptables rules." protocol="IPv6" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.898922 5109 status_manager.go:230] "Starting to sync pod status with apiserver" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.898954 5109 watchdog_linux.go:127] "Systemd watchdog is not enabled or the interval is invalid, so health checking will not be started." Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.898964 5109 kubelet.go:2451] "Starting kubelet main sync loop" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.899009 5109 kubelet.go:2475] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 11 16:52:42 crc kubenswrapper[5109]: W1211 16:52:42.900373 5109 watcher.go:93] Error while processing event ("/sys/fs/cgroup/system.slice/ocp-wait-apiservices-available.service": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/system.slice/ocp-wait-apiservices-available.service: no such file or directory Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.901841 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.906150 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.907664 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.907690 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.907701 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.908551 5109 cpu_manager.go:222] "Starting CPU manager" policy="none" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.908570 5109 cpu_manager.go:223] "Reconciling" reconcilePeriod="10s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.908590 5109 state_mem.go:36] "Initialized new in-memory state store" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.916009 5109 policy_none.go:49] "None policy: Start" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.916029 5109 memory_manager.go:186] "Starting memorymanager" policy="None" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.916043 5109 state_mem.go:35] "Initializing new in-memory state store" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.924873 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.964929 5109 manager.go:341] "Starting Device Plugin manager" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.965167 5109 manager.go:517] "Failed to read data from checkpoint" err="checkpoint is not found" checkpoint="kubelet_internal_checkpoint" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.965187 5109 server.go:85] "Starting device plugin registration server" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.965622 5109 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.965635 5109 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.965870 5109 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.965934 5109 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 11 16:52:42 crc kubenswrapper[5109]: I1211 16:52:42.965941 5109 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.970455 5109 eviction_manager.go:267] "eviction manager: failed to check if we have separate container filesystem. Ignoring." err="non-existent label \"crio-containers\"" Dec 11 16:52:42 crc kubenswrapper[5109]: E1211 16:52:42.970493 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:42.999456 5109 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:42.999701 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.001695 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.001769 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.001787 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.002661 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.002955 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.003037 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.003505 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.003542 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.003556 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.003831 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.003865 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.003875 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.004295 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.004463 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.004522 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.005179 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.005202 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.005211 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.005489 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.005524 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.005542 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.005998 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006122 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006172 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006488 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006517 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006532 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006891 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006924 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.006943 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.007298 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.007425 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.007461 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.007992 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.008019 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.008032 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.008637 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.008664 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.008676 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.008698 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.008724 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.009292 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.009346 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.009368 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.026808 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.041391 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.056773 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.063865 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.065796 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.066862 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.066906 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.066919 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.066944 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.067475 5109 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.079412 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.081338 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.081395 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.081429 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.081456 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.081489 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.088638 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.183829 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.183990 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184030 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184060 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184095 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184154 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184253 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184332 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184360 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184394 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184421 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-run-kubernetes\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184443 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184460 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184476 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184493 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184514 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184532 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184550 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.184946 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-run-kubernetes\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-var-run-kubernetes\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.185037 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-tmp-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.185476 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.185704 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.185708 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-dir\" (UniqueName: \"kubernetes.io/empty-dir/9f0bc7fcb0822a2c13eb2d22cd8c0641-ca-trust-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.185780 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/0b638b8f4bb0070e40528db779baf6a2-tmp\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.185949 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/20c5c5b4bed930554494851fe3cb2b2a-tmp-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.268207 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.270110 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.270179 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.270194 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.270224 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.270708 5109 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.285889 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.285929 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.285947 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.285966 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.285983 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.285997 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286025 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286046 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286061 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286077 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286094 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286107 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286122 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286137 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286156 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286171 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286404 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286448 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-data-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286468 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-usr-local-bin\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286489 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286510 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-log-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286531 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-auto-backup-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-etcd-auto-backup-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286566 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286586 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286606 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/9f0bc7fcb0822a2c13eb2d22cd8c0641-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"9f0bc7fcb0822a2c13eb2d22cd8c0641\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286625 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286645 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-resource-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286663 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-cert-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286680 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286710 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/0b638b8f4bb0070e40528db779baf6a2-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"0b638b8f4bb0070e40528db779baf6a2\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286730 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4e08c320b1e9e2405e6e0107bdf7eeb4-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"4e08c320b1e9e2405e6e0107bdf7eeb4\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.286779 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/20c5c5b4bed930554494851fe3cb2b2a-static-pod-dir\") pod \"etcd-crc\" (UID: \"20c5c5b4bed930554494851fe3cb2b2a\") " pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.342799 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.357982 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.376347 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.380604 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.428107 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.490789 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:52:43 crc kubenswrapper[5109]: W1211 16:52:43.497165 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b638b8f4bb0070e40528db779baf6a2.slice/crio-887b4eb9b479ef502bb4ac90307b637b1c6a315ea6314d88ffec91f4120c7032 WatchSource:0}: Error finding container 887b4eb9b479ef502bb4ac90307b637b1c6a315ea6314d88ffec91f4120c7032: Status 404 returned error can't find the container with id 887b4eb9b479ef502bb4ac90307b637b1c6a315ea6314d88ffec91f4120c7032 Dec 11 16:52:43 crc kubenswrapper[5109]: W1211 16:52:43.499685 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20c5c5b4bed930554494851fe3cb2b2a.slice/crio-a6db6fcedfac74ae1019b4ff52019c3de48c00a8aa9c16b25fb0f8f3fd4b5027 WatchSource:0}: Error finding container a6db6fcedfac74ae1019b4ff52019c3de48c00a8aa9c16b25fb0f8f3fd4b5027: Status 404 returned error can't find the container with id a6db6fcedfac74ae1019b4ff52019c3de48c00a8aa9c16b25fb0f8f3fd4b5027 Dec 11 16:52:43 crc kubenswrapper[5109]: W1211 16:52:43.501589 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3a14caf222afb62aaabdc47808b6f944.slice/crio-d4aa763dcb9f6509bd215021ffab981702b6e4c498d68d67e66b78f8c24b93d4 WatchSource:0}: Error finding container d4aa763dcb9f6509bd215021ffab981702b6e4c498d68d67e66b78f8c24b93d4: Status 404 returned error can't find the container with id d4aa763dcb9f6509bd215021ffab981702b6e4c498d68d67e66b78f8c24b93d4 Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.505079 5109 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 16:52:43 crc kubenswrapper[5109]: W1211 16:52:43.506450 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f0bc7fcb0822a2c13eb2d22cd8c0641.slice/crio-47cd4b60c60c82c2375d31a2a1aefdffc1f3b3c68e1e046d2e7b509cdb4468e0 WatchSource:0}: Error finding container 47cd4b60c60c82c2375d31a2a1aefdffc1f3b3c68e1e046d2e7b509cdb4468e0: Status 404 returned error can't find the container with id 47cd4b60c60c82c2375d31a2a1aefdffc1f3b3c68e1e046d2e7b509cdb4468e0 Dec 11 16:52:43 crc kubenswrapper[5109]: W1211 16:52:43.516404 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4e08c320b1e9e2405e6e0107bdf7eeb4.slice/crio-e6604b7745465243a68ebc5f377a7729efb00d39a83ea3c9770f02f323e0cfb4 WatchSource:0}: Error finding container e6604b7745465243a68ebc5f377a7729efb00d39a83ea3c9770f02f323e0cfb4: Status 404 returned error can't find the container with id e6604b7745465243a68ebc5f377a7729efb00d39a83ea3c9770f02f323e0cfb4 Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.671262 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.673872 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.673926 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.673939 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.673965 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.674472 5109 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.808295 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 11 16:52:43 crc kubenswrapper[5109]: E1211 16:52:43.835343 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.910322 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"a6db6fcedfac74ae1019b4ff52019c3de48c00a8aa9c16b25fb0f8f3fd4b5027"} Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.912957 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"887b4eb9b479ef502bb4ac90307b637b1c6a315ea6314d88ffec91f4120c7032"} Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.915548 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"47cd4b60c60c82c2375d31a2a1aefdffc1f3b3c68e1e046d2e7b509cdb4468e0"} Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.918121 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerStarted","Data":"e6604b7745465243a68ebc5f377a7729efb00d39a83ea3c9770f02f323e0cfb4"} Dec 11 16:52:43 crc kubenswrapper[5109]: I1211 16:52:43.927093 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"d4aa763dcb9f6509bd215021ffab981702b6e4c498d68d67e66b78f8c24b93d4"} Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.113246 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.153662 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.229459 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.332070 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.475294 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.476635 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.476695 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.476717 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.476784 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.477468 5109 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.802169 5109 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.803263 5109 certificate_manager.go:596] "Failed while requesting a signed certificate from the control plane" err="cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="kubernetes.io/kube-apiserver-client-kubelet.UnhandledError" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.808083 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.932442 5109 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd" exitCode=0 Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.932557 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd"} Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.932710 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.933844 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.933882 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.933893 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.934268 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.935925 5109 generic.go:358] "Generic (PLEG): container finished" podID="0b638b8f4bb0070e40528db779baf6a2" containerID="f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3" exitCode=0 Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.935987 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerDied","Data":"f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3"} Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.936180 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.936914 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.936958 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.936974 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.937188 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.939535 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"58b2adf4e750dc18bc35c0aea18178cf6daed481b29ea518459defa47769e5d0"} Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.940196 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"b9608c14d8ea51cb0920addf885ee63ff3e4c8a81ab46a226edebd9a0bb6a399"} Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.940227 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6"} Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.946522 5109 generic.go:358] "Generic (PLEG): container finished" podID="4e08c320b1e9e2405e6e0107bdf7eeb4" containerID="e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014" exitCode=0 Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.946653 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerDied","Data":"e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014"} Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.946890 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.949117 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.949206 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.949275 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.949500 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.952444 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a" exitCode=0 Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.952523 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a"} Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.952727 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.953616 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.953666 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.953684 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:44 crc kubenswrapper[5109]: E1211 16:52:44.953953 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.955865 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.956423 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.956515 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:44 crc kubenswrapper[5109]: I1211 16:52:44.956583 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.811214 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.17:6443: connect: connection refused Dec 11 16:52:45 crc kubenswrapper[5109]: E1211 16:52:45.833059 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="3.2s" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.963418 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.963470 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.963484 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.965427 5109 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf" exitCode=0 Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.965511 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.965678 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.966548 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.966581 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.966607 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:45 crc kubenswrapper[5109]: E1211 16:52:45.966823 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.971173 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"c37b01eb1234360d043eff79f1eb374f5d95c1132e416be3c05ad3fd9fdfc36c"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.971204 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"b0e470f587ad2b1f84cb1100ce6155dbe9ba3e4a36c01970e843d5cdf38b3d56"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.971216 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"0b638b8f4bb0070e40528db779baf6a2","Type":"ContainerStarted","Data":"c10e7a3d56908a1837ca1b29a944fbe9febec4030a733cd53c0ecda0177091c9"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.971364 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.971966 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.972002 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.972014 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:45 crc kubenswrapper[5109]: E1211 16:52:45.972222 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.975733 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"1cbeee4ade3904eb92ab061e705d6387d66e4f501d3ee96e67260accf10d2acb"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.975889 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.976379 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.976414 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.976427 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:45 crc kubenswrapper[5109]: E1211 16:52:45.976617 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.978600 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"4e08c320b1e9e2405e6e0107bdf7eeb4","Type":"ContainerStarted","Data":"6ab4918fcfdbfaadb2441c5d160818ac2fd2f2d718e73fc0ef7d9d62dc004b11"} Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.978670 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.979304 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.979341 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:45 crc kubenswrapper[5109]: I1211 16:52:45.979354 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:45 crc kubenswrapper[5109]: E1211 16:52:45.979515 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.068822 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.077744 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.078600 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.078659 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.078671 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.078696 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.079121 5109 kubelet_node_status.go:110] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.17:6443: connect: connection refused" node="crc" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.192762 5109 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.1880376574b07590 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.818180496 +0000 UTC m=+0.497871972,LastTimestamp:2025-12-11 16:52:42.818180496 +0000 UTC m=+0.497871972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.271474 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.17:6443: connect: connection refused" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.985034 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"cba673008bae8a23835d00ab68ce3e08f89b0e9bb926a5fbfbc41e522337c803"} Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.985074 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a"} Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.985191 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.985775 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.985803 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.985815 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.985975 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.987932 5109 generic.go:358] "Generic (PLEG): container finished" podID="20c5c5b4bed930554494851fe3cb2b2a" containerID="35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf" exitCode=0 Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.988020 5109 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.988044 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.988214 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.988319 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerDied","Data":"35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf"} Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.988413 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.988721 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989127 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989148 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989156 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.989314 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989503 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989522 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989530 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.989694 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989905 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989933 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.989941 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.990065 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.990283 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.990304 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:46 crc kubenswrapper[5109]: I1211 16:52:46.990312 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:46 crc kubenswrapper[5109]: E1211 16:52:46.990481 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.641874 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.848964 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.995774 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"352c8d599c93e387e36a5488dcc02f39161e4d8c0a9fbf050e4e19c3c1a4fd92"} Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.995819 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"0ec3b28b6104ac258b2e97c9f10ec3d55ea93bbf9d9165fdce21648201c85e9c"} Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.995832 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"a3b479ab0d924d9fbb549995801bad80a45ec723aff9b99d07b5d67b37675cc6"} Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.995945 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.995965 5109 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.996072 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.997346 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.997429 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.997466 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:47 crc kubenswrapper[5109]: E1211 16:52:47.998084 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.999167 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.999235 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:47 crc kubenswrapper[5109]: I1211 16:52:47.999260 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:47 crc kubenswrapper[5109]: E1211 16:52:47.999885 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:48 crc kubenswrapper[5109]: I1211 16:52:48.935278 5109 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kube-apiserver-client-kubelet" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.005179 5109 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.005257 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.005721 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.006014 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"ca35d4089afda84920619ba9d77e28fdb8bf432b26bf3cb949151fce931db7a1"} Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.006053 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"20c5c5b4bed930554494851fe3cb2b2a","Type":"ContainerStarted","Data":"445fc06154e726c5f8c28bf3f36d0bc02dde5ccc34cfefcb928cbee1eb9c0d85"} Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.006408 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.006454 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.006473 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:49 crc kubenswrapper[5109]: E1211 16:52:49.006824 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.007850 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.007894 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.007914 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:49 crc kubenswrapper[5109]: E1211 16:52:49.008492 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.279392 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.280760 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.280819 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.280839 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.280887 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.287849 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.288087 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.288897 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.289072 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.289228 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:49 crc kubenswrapper[5109]: E1211 16:52:49.289856 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:49 crc kubenswrapper[5109]: I1211 16:52:49.474542 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.010159 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.010923 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.010988 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.012142 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.012250 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.012264 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.012313 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.012330 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.012282 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:50 crc kubenswrapper[5109]: E1211 16:52:50.012936 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:50 crc kubenswrapper[5109]: E1211 16:52:50.013531 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.372032 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.372370 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.374328 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.374387 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.374408 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:50 crc kubenswrapper[5109]: E1211 16:52:50.374915 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.509415 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.509693 5109 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.509781 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.511080 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.511156 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:50 crc kubenswrapper[5109]: I1211 16:52:50.511174 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:50 crc kubenswrapper[5109]: E1211 16:52:50.511906 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.014137 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.015125 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.015197 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.015217 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:51 crc kubenswrapper[5109]: E1211 16:52:51.015876 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.094485 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.925168 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.925805 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.927015 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.927090 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:51 crc kubenswrapper[5109]: I1211 16:52:51.927110 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:51 crc kubenswrapper[5109]: E1211 16:52:51.927828 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.016936 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.018111 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.018172 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.018187 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:52 crc kubenswrapper[5109]: E1211 16:52:52.018619 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.930989 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.931645 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.933049 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.933148 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:52 crc kubenswrapper[5109]: I1211 16:52:52.933176 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:52 crc kubenswrapper[5109]: E1211 16:52:52.933978 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:52 crc kubenswrapper[5109]: E1211 16:52:52.970880 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:52:53 crc kubenswrapper[5109]: I1211 16:52:53.071907 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-etcd/etcd-crc" Dec 11 16:52:53 crc kubenswrapper[5109]: I1211 16:52:53.072355 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:52:53 crc kubenswrapper[5109]: I1211 16:52:53.073758 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:52:53 crc kubenswrapper[5109]: I1211 16:52:53.073792 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:52:53 crc kubenswrapper[5109]: I1211 16:52:53.073808 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:52:53 crc kubenswrapper[5109]: E1211 16:52:53.074257 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:52:54 crc kubenswrapper[5109]: I1211 16:52:54.094712 5109 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 16:52:54 crc kubenswrapper[5109]: I1211 16:52:54.094840 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 11 16:52:56 crc kubenswrapper[5109]: I1211 16:52:56.809664 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 11 16:52:56 crc kubenswrapper[5109]: I1211 16:52:56.980455 5109 trace.go:236] Trace[327407311]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 16:52:46.979) (total time: 10001ms): Dec 11 16:52:56 crc kubenswrapper[5109]: Trace[327407311]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10000ms (16:52:56.980) Dec 11 16:52:56 crc kubenswrapper[5109]: Trace[327407311]: [10.00101446s] [10.00101446s] END Dec 11 16:52:56 crc kubenswrapper[5109]: E1211 16:52:56.980504 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 11 16:52:57 crc kubenswrapper[5109]: I1211 16:52:57.041650 5109 trace.go:236] Trace[1657634378]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 16:52:47.039) (total time: 10001ms): Dec 11 16:52:57 crc kubenswrapper[5109]: Trace[1657634378]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:52:57.041) Dec 11 16:52:57 crc kubenswrapper[5109]: Trace[1657634378]: [10.00191847s] [10.00191847s] END Dec 11 16:52:57 crc kubenswrapper[5109]: E1211 16:52:57.041691 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 11 16:52:57 crc kubenswrapper[5109]: I1211 16:52:57.523601 5109 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 16:52:57 crc kubenswrapper[5109]: I1211 16:52:57.523670 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 16:52:57 crc kubenswrapper[5109]: I1211 16:52:57.531879 5109 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 16:52:57 crc kubenswrapper[5109]: I1211 16:52:57.531956 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 11 16:52:57 crc kubenswrapper[5109]: I1211 16:52:57.860074 5109 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]log ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]etcd ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/generic-apiserver-start-informers ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/priority-and-fairness-filter ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-apiextensions-informers ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-apiextensions-controllers ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/crd-informer-synced ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-system-namespaces-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 11 16:52:57 crc kubenswrapper[5109]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 11 16:52:57 crc kubenswrapper[5109]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/bootstrap-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-kubernetes-service-cidr-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/start-kube-aggregator-informers ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/apiservice-registration-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/apiservice-discovery-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]autoregister-completion ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/apiservice-openapi-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 11 16:52:57 crc kubenswrapper[5109]: livez check failed Dec 11 16:52:57 crc kubenswrapper[5109]: I1211 16:52:57.860174 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:52:59 crc kubenswrapper[5109]: E1211 16:52:59.036714 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Dec 11 16:53:01 crc kubenswrapper[5109]: E1211 16:53:01.141973 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 11 16:53:01 crc kubenswrapper[5109]: I1211 16:53:01.955721 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 11 16:53:01 crc kubenswrapper[5109]: I1211 16:53:01.956084 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:01 crc kubenswrapper[5109]: I1211 16:53:01.957225 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:01 crc kubenswrapper[5109]: I1211 16:53:01.957263 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:01 crc kubenswrapper[5109]: I1211 16:53:01.957279 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:01 crc kubenswrapper[5109]: E1211 16:53:01.957841 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:01 crc kubenswrapper[5109]: I1211 16:53:01.969718 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.022978 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.023156 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.024245 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.024336 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.024359 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.025213 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.043279 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.043895 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.043923 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.043935 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.044283 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.540549 5109 trace.go:236] Trace[418346290]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 16:52:51.536) (total time: 11003ms): Dec 11 16:53:02 crc kubenswrapper[5109]: Trace[418346290]: ---"Objects listed" error:nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope 11003ms (16:53:02.540) Dec 11 16:53:02 crc kubenswrapper[5109]: Trace[418346290]: [11.003634859s] [11.003634859s] END Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.540596 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.540686 5109 trace.go:236] Trace[1344552428]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Dec-2025 16:52:49.410) (total time: 13129ms): Dec 11 16:53:02 crc kubenswrapper[5109]: Trace[1344552428]: ---"Objects listed" error:runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope 13129ms (16:53:02.540) Dec 11 16:53:02 crc kubenswrapper[5109]: Trace[1344552428]: [13.129873159s] [13.129873159s] END Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.540605 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.1880376574b07590 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.818180496 +0000 UTC m=+0.497871972,LastTimestamp:2025-12-11 16:52:42.818180496 +0000 UTC m=+0.497871972,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.540706 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.551983 5109 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.552155 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.558080 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.559054 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0682b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,LastTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.562200 5109 reflector.go:430] "Caches populated" logger="kubernetes.io/kube-apiserver-client-kubelet" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.570894 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657d9f59fc default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.968054268 +0000 UTC m=+0.647745734,LastTimestamp:2025-12-11 16:52:42.968054268 +0000 UTC m=+0.647745734,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.591392 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a061c34\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:43.001727924 +0000 UTC m=+0.681419390,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.600120 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0659e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:43.001778231 +0000 UTC m=+0.681469707,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.609691 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0682b2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0682b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,LastTimestamp:2025-12-11 16:52:43.00179327 +0000 UTC m=+0.681484736,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.617056 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a061c34\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:43.003526062 +0000 UTC m=+0.683217528,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.622221 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0659e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:43.003549931 +0000 UTC m=+0.683241407,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.627116 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0682b2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0682b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,LastTimestamp:2025-12-11 16:52:43.00356145 +0000 UTC m=+0.683252916,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.630834 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a061c34\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:43.003847743 +0000 UTC m=+0.683539209,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.635880 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0659e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:43.003870772 +0000 UTC m=+0.683562238,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.639652 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0682b2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0682b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,LastTimestamp:2025-12-11 16:52:43.003880181 +0000 UTC m=+0.683571647,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.644793 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a061c34\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:43.00519529 +0000 UTC m=+0.684886756,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.648307 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0659e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:43.005207289 +0000 UTC m=+0.684898755,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.651909 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0682b2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0682b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,LastTimestamp:2025-12-11 16:52:43.005215868 +0000 UTC m=+0.684907334,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.656973 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a061c34\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:43.00550911 +0000 UTC m=+0.685200606,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.662280 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0659e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:43.005534279 +0000 UTC m=+0.685225775,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.667996 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0682b2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0682b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,LastTimestamp:2025-12-11 16:52:43.005551287 +0000 UTC m=+0.685242793,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.673592 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a061c34\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:43.006508049 +0000 UTC m=+0.686199515,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.682836 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0659e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:43.006524748 +0000 UTC m=+0.686216214,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.688093 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0682b2\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0682b2 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907706034 +0000 UTC m=+0.587397500,LastTimestamp:2025-12-11 16:52:43.006538017 +0000 UTC m=+0.686229483,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.695275 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a061c34\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a061c34 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907679796 +0000 UTC m=+0.587371262,LastTimestamp:2025-12-11 16:52:43.006914173 +0000 UTC m=+0.686605659,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.701435 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.188037657a0659e1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.188037657a0659e1 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:42.907695585 +0000 UTC m=+0.587387051,LastTimestamp:2025-12-11 16:52:43.006937852 +0000 UTC m=+0.686629328,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.713055 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.188037659da7af77 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:43.505471351 +0000 UTC m=+1.185162817,LastTimestamp:2025-12-11 16:52:43.505471351 +0000 UTC m=+1.185162817,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.717860 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.188037659da99ad7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:43.505597143 +0000 UTC m=+1.185288649,LastTimestamp:2025-12-11 16:52:43.505597143 +0000 UTC m=+1.185288649,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.722034 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037659daf517c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:43.50597158 +0000 UTC m=+1.185663046,LastTimestamp:2025-12-11 16:52:43.50597158 +0000 UTC m=+1.185663046,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.725964 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.188037659e47ce99 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:43.515965081 +0000 UTC m=+1.195656547,LastTimestamp:2025-12-11 16:52:43.515965081 +0000 UTC m=+1.195656547,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.731017 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.188037659e707b96 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:43.518630806 +0000 UTC m=+1.198322292,LastTimestamp:2025-12-11 16:52:43.518630806 +0000 UTC m=+1.198322292,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.736061 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765bfe00f3f openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container: kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.079591231 +0000 UTC m=+1.759282697,LastTimestamp:2025-12-11 16:52:44.079591231 +0000 UTC m=+1.759282697,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.743481 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803765bfe1b9d3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.079700435 +0000 UTC m=+1.759391891,LastTimestamp:2025-12-11 16:52:44.079700435 +0000 UTC m=+1.759391891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.750636 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18803765bfe21de5 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container: wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.079726053 +0000 UTC m=+1.759417519,LastTimestamp:2025-12-11 16:52:44.079726053 +0000 UTC m=+1.759417519,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.753681 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803765bfe1b785 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.079699845 +0000 UTC m=+1.759391351,LastTimestamp:2025-12-11 16:52:44.079699845 +0000 UTC m=+1.759391351,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.757168 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18803765c07fc3c9 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container: setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.090057673 +0000 UTC m=+1.769749139,LastTimestamp:2025-12-11 16:52:44.090057673 +0000 UTC m=+1.769749139,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.762310 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18803765c0bc9c48 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.094045256 +0000 UTC m=+1.773736762,LastTimestamp:2025-12-11 16:52:44.094045256 +0000 UTC m=+1.773736762,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.769933 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765c0bdc38a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.094120842 +0000 UTC m=+1.773812308,LastTimestamp:2025-12-11 16:52:44.094120842 +0000 UTC m=+1.773812308,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.774204 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803765c0c08583 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.094301571 +0000 UTC m=+1.773993037,LastTimestamp:2025-12-11 16:52:44.094301571 +0000 UTC m=+1.773993037,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.779642 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765c0d126d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.095391443 +0000 UTC m=+1.775082909,LastTimestamp:2025-12-11 16:52:44.095391443 +0000 UTC m=+1.775082909,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.784470 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803765c109d4d3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.099106003 +0000 UTC m=+1.778797459,LastTimestamp:2025-12-11 16:52:44.099106003 +0000 UTC m=+1.778797459,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.790476 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18803765c155364d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.104046157 +0000 UTC m=+1.783737623,LastTimestamp:2025-12-11 16:52:44.104046157 +0000 UTC m=+1.783737623,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.797578 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765d54c558c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container: cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.439008652 +0000 UTC m=+2.118700158,LastTimestamp:2025-12-11 16:52:44.439008652 +0000 UTC m=+2.118700158,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.809872 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765d617f6d3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.452353747 +0000 UTC m=+2.132045243,LastTimestamp:2025-12-11 16:52:44.452353747 +0000 UTC m=+2.132045243,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.810205 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.814585 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765d6350464 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.454257764 +0000 UTC m=+2.133949270,LastTimestamp:2025-12-11 16:52:44.454257764 +0000 UTC m=+2.133949270,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.820075 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765ee806e07 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container: kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.861853191 +0000 UTC m=+2.541544657,LastTimestamp:2025-12-11 16:52:44.861853191 +0000 UTC m=+2.541544657,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.823707 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765ef31fb77 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.873489271 +0000 UTC m=+2.553180737,LastTimestamp:2025-12-11 16:52:44.873489271 +0000 UTC m=+2.553180737,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.827823 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.18803765ef448c3a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.874705978 +0000 UTC m=+2.554397444,LastTimestamp:2025-12-11 16:52:44.874705978 +0000 UTC m=+2.554397444,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.831819 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803765f2ef8b52 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.93624405 +0000 UTC m=+2.615935546,LastTimestamp:2025-12-11 16:52:44.93624405 +0000 UTC m=+2.615935546,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.835780 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18803765f30a79e8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.938009064 +0000 UTC m=+2.617700540,LastTimestamp:2025-12-11 16:52:44.938009064 +0000 UTC m=+2.617700540,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.839578 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.18803765f3f76c84 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.953537668 +0000 UTC m=+2.633229134,LastTimestamp:2025-12-11 16:52:44.953537668 +0000 UTC m=+2.633229134,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.843475 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803765f418d02d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:44.955725869 +0000 UTC m=+2.635417335,LastTimestamp:2025-12-11 16:52:44.955725869 +0000 UTC m=+2.635417335,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.847191 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.188037660307cd8c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container: kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.206269324 +0000 UTC m=+2.885960790,LastTimestamp:2025-12-11 16:52:45.206269324 +0000 UTC m=+2.885960790,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.851115 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.188037660400ab05 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.222578949 +0000 UTC m=+2.902270415,LastTimestamp:2025-12-11 16:52:45.222578949 +0000 UTC m=+2.902270415,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.854181 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1880376606a3a08f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container: kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.266813071 +0000 UTC m=+2.946504787,LastTimestamp:2025-12-11 16:52:45.266813071 +0000 UTC m=+2.946504787,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.854967 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.855156 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.855515 5109 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.855570 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.856029 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.856077 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.856088 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.856426 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.858796 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376607b8aab7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container: kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.284969143 +0000 UTC m=+2.964660599,LastTimestamp:2025-12-11 16:52:45.284969143 +0000 UTC m=+2.964660599,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.859410 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.863305 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1880376607bb476f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.285140335 +0000 UTC m=+2.964831811,LastTimestamp:2025-12-11 16:52:45.285140335 +0000 UTC m=+2.964831811,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.867428 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1880376607ccdc29 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container: etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.286292521 +0000 UTC m=+2.965983987,LastTimestamp:2025-12-11 16:52:45.286292521 +0000 UTC m=+2.965983987,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.871193 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1880376607d29fe2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.286670306 +0000 UTC m=+2.966361772,LastTimestamp:2025-12-11 16:52:45.286670306 +0000 UTC m=+2.966361772,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.875781 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1880376607d5d147 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container: kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.286879559 +0000 UTC m=+2.966571015,LastTimestamp:2025-12-11 16:52:45.286879559 +0000 UTC m=+2.966571015,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.879474 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.1880376608fc4990 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:4e08c320b1e9e2405e6e0107bdf7eeb4,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.306177936 +0000 UTC m=+2.985869402,LastTimestamp:2025-12-11 16:52:45.306177936 +0000 UTC m=+2.985869402,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.884187 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037660965ca5f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.313092191 +0000 UTC m=+2.992783657,LastTimestamp:2025-12-11 16:52:45.313092191 +0000 UTC m=+2.992783657,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.889420 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376609661ca4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.313113252 +0000 UTC m=+2.992804738,LastTimestamp:2025-12-11 16:52:45.313113252 +0000 UTC m=+2.992804738,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.895074 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376609ae31c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.317837257 +0000 UTC m=+2.997528723,LastTimestamp:2025-12-11 16:52:45.317837257 +0000 UTC m=+2.997528723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.900244 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18803766163135b6 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container: kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.52775007 +0000 UTC m=+3.207441536,LastTimestamp:2025-12-11 16:52:45.52775007 +0000 UTC m=+3.207441536,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.906345 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.188037661665f6b4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container: kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.531207348 +0000 UTC m=+3.210898814,LastTimestamp:2025-12-11 16:52:45.531207348 +0000 UTC m=+3.210898814,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.911220 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18803766178b4dd9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.550431705 +0000 UTC m=+3.230123171,LastTimestamp:2025-12-11 16:52:45.550431705 +0000 UTC m=+3.230123171,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.915132 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.18803766179b4a66 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.551479398 +0000 UTC m=+3.231170864,LastTimestamp:2025-12-11 16:52:45.551479398 +0000 UTC m=+3.231170864,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.919716 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376617d41fdc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.55520406 +0000 UTC m=+3.234895526,LastTimestamp:2025-12-11 16:52:45.55520406 +0000 UTC m=+3.234895526,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.924485 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376617eca9be openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.556812222 +0000 UTC m=+3.236503688,LastTimestamp:2025-12-11 16:52:45.556812222 +0000 UTC m=+3.236503688,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.929219 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1880376625262aaf openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container: kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.778684591 +0000 UTC m=+3.458376057,LastTimestamp:2025-12-11 16:52:45.778684591 +0000 UTC m=+3.458376057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.933310 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.933587 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.934266 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803766253d3a8e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container: kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.780195982 +0000 UTC m=+3.459887448,LastTimestamp:2025-12-11 16:52:45.780195982 +0000 UTC m=+3.459887448,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.934502 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.934551 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.934567 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.935026 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.939435 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376626415b78 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.797243768 +0000 UTC m=+3.476935234,LastTimestamp:2025-12-11 16:52:45.797243768 +0000 UTC m=+3.476935234,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: I1211 16:53:02.940568 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.944633 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803766265fdbef openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.799242735 +0000 UTC m=+3.478934201,LastTimestamp:2025-12-11 16:52:45.799242735 +0000 UTC m=+3.478934201,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.951184 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.1880376627b8004b openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:0b638b8f4bb0070e40528db779baf6a2,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.821796427 +0000 UTC m=+3.501487933,LastTimestamp:2025-12-11 16:52:45.821796427 +0000 UTC m=+3.501487933,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.956522 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803766308ee456 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:45.970097238 +0000 UTC m=+3.649788714,LastTimestamp:2025-12-11 16:52:45.970097238 +0000 UTC m=+3.649788714,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.965542 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376633f604ee openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container: kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.027187438 +0000 UTC m=+3.706878904,LastTimestamp:2025-12-11 16:52:46.027187438 +0000 UTC m=+3.706878904,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.970912 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803766353b1078 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.048489592 +0000 UTC m=+3.728181048,LastTimestamp:2025-12-11 16:52:46.048489592 +0000 UTC m=+3.728181048,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.971158 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.976102 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376635634f63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.051127139 +0000 UTC m=+3.730818605,LastTimestamp:2025-12-11 16:52:46.051127139 +0000 UTC m=+3.730818605,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.981574 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.1880376640ab35b8 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container: etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.240388536 +0000 UTC m=+3.920080012,LastTimestamp:2025-12-11 16:52:46.240388536 +0000 UTC m=+3.920080012,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.985711 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803766425ab2f9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.268666617 +0000 UTC m=+3.948358083,LastTimestamp:2025-12-11 16:52:46.268666617 +0000 UTC m=+3.948358083,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.989800 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376643e494f3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.294480115 +0000 UTC m=+3.974171591,LastTimestamp:2025-12-11 16:52:46.294480115 +0000 UTC m=+3.974171591,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.994163 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376644ca901a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.309552154 +0000 UTC m=+3.989243630,LastTimestamp:2025-12-11 16:52:46.309552154 +0000 UTC m=+3.989243630,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:02 crc kubenswrapper[5109]: E1211 16:53:02.999276 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037666d6f8935 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.991452469 +0000 UTC m=+4.671143935,LastTimestamp:2025-12-11 16:52:46.991452469 +0000 UTC m=+4.671143935,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.005113 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037667ce7e78e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container: etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.250999182 +0000 UTC m=+4.930690658,LastTimestamp:2025-12-11 16:52:47.250999182 +0000 UTC m=+4.930690658,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.009692 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037667dbce942 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.264958786 +0000 UTC m=+4.944650262,LastTimestamp:2025-12-11 16:52:47.264958786 +0000 UTC m=+4.944650262,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.013920 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037667dd12ad5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.266286293 +0000 UTC m=+4.945977769,LastTimestamp:2025-12-11 16:52:47.266286293 +0000 UTC m=+4.945977769,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.018620 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037668bc48268 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container: etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.500337768 +0000 UTC m=+5.180029234,LastTimestamp:2025-12-11 16:52:47.500337768 +0000 UTC m=+5.180029234,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.024661 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037668cad7f90 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.515606928 +0000 UTC m=+5.195298394,LastTimestamp:2025-12-11 16:52:47.515606928 +0000 UTC m=+5.195298394,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.028303 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037668cbdff88 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.516688264 +0000 UTC m=+5.196379730,LastTimestamp:2025-12-11 16:52:47.516688264 +0000 UTC m=+5.196379730,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.033058 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037669d00d681 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container: etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.789504129 +0000 UTC m=+5.469195625,LastTimestamp:2025-12-11 16:52:47.789504129 +0000 UTC m=+5.469195625,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.037871 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037669e349180 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.809671552 +0000 UTC m=+5.489363048,LastTimestamp:2025-12-11 16:52:47.809671552 +0000 UTC m=+5.489363048,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.042234 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.188037669e4cb681 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:47.811253889 +0000 UTC m=+5.490945365,LastTimestamp:2025-12-11 16:52:47.811253889 +0000 UTC m=+5.490945365,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.046178 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.048936 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803766ae2ac5d3 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container: etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:48.077465043 +0000 UTC m=+5.757156549,LastTimestamp:2025-12-11 16:52:48.077465043 +0000 UTC m=+5.757156549,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.049130 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="cba673008bae8a23835d00ab68ce3e08f89b0e9bb926a5fbfbc41e522337c803" exitCode=255 Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.049362 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.049427 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"cba673008bae8a23835d00ab68ce3e08f89b0e9bb926a5fbfbc41e522337c803"} Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.049539 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.050211 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.050304 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.050367 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.051340 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.051670 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.051720 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.051773 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.051976 5109 scope.go:117] "RemoveContainer" containerID="cba673008bae8a23835d00ab68ce3e08f89b0e9bb926a5fbfbc41e522337c803" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.052068 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.053292 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803766af22f943 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:48.093731139 +0000 UTC m=+5.773422635,LastTimestamp:2025-12-11 16:52:48.093731139 +0000 UTC m=+5.773422635,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.060494 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803766af43ff17 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:48.095895319 +0000 UTC m=+5.775586825,LastTimestamp:2025-12-11 16:52:48.095895319 +0000 UTC m=+5.775586825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.064836 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803766bf13c30f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container: etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:48.361169679 +0000 UTC m=+6.040861185,LastTimestamp:2025-12-11 16:52:48.361169679 +0000 UTC m=+6.040861185,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.077268 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.18803766c0455d3f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:20c5c5b4bed930554494851fe3cb2b2a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:48.381197631 +0000 UTC m=+6.060889137,LastTimestamp:2025-12-11 16:52:48.381197631 +0000 UTC m=+6.060889137,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.085824 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Dec 11 16:53:03 crc kubenswrapper[5109]: &Event{ObjectMeta:{kube-controller-manager-crc.1880376814d44688 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://localhost:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Dec 11 16:53:03 crc kubenswrapper[5109]: body: Dec 11 16:53:03 crc kubenswrapper[5109]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:54.094816904 +0000 UTC m=+11.774508390,LastTimestamp:2025-12-11 16:52:54.094816904 +0000 UTC m=+11.774508390,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 11 16:53:03 crc kubenswrapper[5109]: > Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.090120 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.1880376814d57e38 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:9f0bc7fcb0822a2c13eb2d22cd8c0641,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:54.094896696 +0000 UTC m=+11.774588162,LastTimestamp:2025-12-11 16:52:54.094896696 +0000 UTC m=+11.774588162,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.095266 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Dec 11 16:53:03 crc kubenswrapper[5109]: &Event{ObjectMeta:{kube-apiserver-crc.18803768e133ffe6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Dec 11 16:53:03 crc kubenswrapper[5109]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 16:53:03 crc kubenswrapper[5109]: Dec 11 16:53:03 crc kubenswrapper[5109]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:57.523642342 +0000 UTC m=+15.203333818,LastTimestamp:2025-12-11 16:52:57.523642342 +0000 UTC m=+15.203333818,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 11 16:53:03 crc kubenswrapper[5109]: > Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.098903 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803768e134c2c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:57.523692233 +0000 UTC m=+15.203383709,LastTimestamp:2025-12-11 16:52:57.523692233 +0000 UTC m=+15.203383709,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.104229 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18803768e133ffe6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Dec 11 16:53:03 crc kubenswrapper[5109]: &Event{ObjectMeta:{kube-apiserver-crc.18803768e133ffe6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Dec 11 16:53:03 crc kubenswrapper[5109]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 11 16:53:03 crc kubenswrapper[5109]: Dec 11 16:53:03 crc kubenswrapper[5109]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:57.523642342 +0000 UTC m=+15.203333818,LastTimestamp:2025-12-11 16:52:57.531928911 +0000 UTC m=+15.211620387,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 11 16:53:03 crc kubenswrapper[5109]: > Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.108247 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.18803768e134c2c9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803768e134c2c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:57.523692233 +0000 UTC m=+15.203383709,LastTimestamp:2025-12-11 16:52:57.531989632 +0000 UTC m=+15.211681118,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.114159 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Dec 11 16:53:03 crc kubenswrapper[5109]: &Event{ObjectMeta:{kube-apiserver-crc.18803768f5429df5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 500 Dec 11 16:53:03 crc kubenswrapper[5109]: body: [+]ping ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]log ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]etcd ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/generic-apiserver-start-informers ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/priority-and-fairness-filter ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-apiextensions-informers ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-apiextensions-controllers ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/crd-informer-synced ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-system-namespaces-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 11 16:53:03 crc kubenswrapper[5109]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 11 16:53:03 crc kubenswrapper[5109]: [-]poststarthook/scheduling/bootstrap-system-priority-classes failed: reason withheld Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/bootstrap-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-kubernetes-service-cidr-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/start-kube-aggregator-informers ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/apiservice-registration-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/apiservice-discovery-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]autoregister-completion ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/apiservice-openapi-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 11 16:53:03 crc kubenswrapper[5109]: livez check failed Dec 11 16:53:03 crc kubenswrapper[5109]: Dec 11 16:53:03 crc kubenswrapper[5109]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:57.860144629 +0000 UTC m=+15.539836135,LastTimestamp:2025-12-11 16:52:57.860144629 +0000 UTC m=+15.539836135,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 11 16:53:03 crc kubenswrapper[5109]: > Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.118686 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.18803768f54385cd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 500,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:57.860203981 +0000 UTC m=+15.539895487,LastTimestamp:2025-12-11 16:52:57.860203981 +0000 UTC m=+15.539895487,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.124523 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Dec 11 16:53:03 crc kubenswrapper[5109]: &Event{ObjectMeta:{kube-apiserver-crc.1880376a1f026c31 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:ProbeError,Message:Readiness probe error: Get "https://192.168.126.11:17697/healthz": dial tcp 192.168.126.11:17697: connect: connection refused Dec 11 16:53:03 crc kubenswrapper[5109]: body: Dec 11 16:53:03 crc kubenswrapper[5109]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:02.855547953 +0000 UTC m=+20.535239419,LastTimestamp:2025-12-11 16:53:02.855547953 +0000 UTC m=+20.535239419,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Dec 11 16:53:03 crc kubenswrapper[5109]: > Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.129096 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376a1f04585a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Unhealthy,Message:Readiness probe failed: Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:02.855673946 +0000 UTC m=+20.535365412,LastTimestamp:2025-12-11 16:53:02.855673946 +0000 UTC m=+20.535365412,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.134957 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376635634f63\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376635634f63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.051127139 +0000 UTC m=+3.730818605,LastTimestamp:2025-12-11 16:53:03.0532882 +0000 UTC m=+20.732979666,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.241797 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376643e494f3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376643e494f3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.294480115 +0000 UTC m=+3.974171591,LastTimestamp:2025-12-11 16:53:03.237247556 +0000 UTC m=+20.916939022,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.245454 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 11 16:53:03 crc kubenswrapper[5109]: E1211 16:53:03.251142 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376644ca901a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376644ca901a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.309552154 +0000 UTC m=+3.989243630,LastTimestamp:2025-12-11 16:53:03.246540779 +0000 UTC m=+20.926232235,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:03 crc kubenswrapper[5109]: I1211 16:53:03.813365 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:04 crc kubenswrapper[5109]: I1211 16:53:04.052571 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Dec 11 16:53:04 crc kubenswrapper[5109]: I1211 16:53:04.053954 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d"} Dec 11 16:53:04 crc kubenswrapper[5109]: I1211 16:53:04.054075 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:04 crc kubenswrapper[5109]: I1211 16:53:04.054492 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:04 crc kubenswrapper[5109]: I1211 16:53:04.054524 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:04 crc kubenswrapper[5109]: I1211 16:53:04.054534 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:04 crc kubenswrapper[5109]: E1211 16:53:04.054846 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:04 crc kubenswrapper[5109]: I1211 16:53:04.820063 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.057735 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.058220 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/0.log" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.059976 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d" exitCode=255 Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.060032 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d"} Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.060079 5109 scope.go:117] "RemoveContainer" containerID="cba673008bae8a23835d00ab68ce3e08f89b0e9bb926a5fbfbc41e522337c803" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.060309 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.061199 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.061233 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.061247 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:05 crc kubenswrapper[5109]: E1211 16:53:05.061560 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.061920 5109 scope.go:117] "RemoveContainer" containerID="8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d" Dec 11 16:53:05 crc kubenswrapper[5109]: E1211 16:53:05.062199 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:05 crc kubenswrapper[5109]: E1211 16:53:05.075998 5109 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:05 crc kubenswrapper[5109]: E1211 16:53:05.446312 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 11 16:53:05 crc kubenswrapper[5109]: I1211 16:53:05.814870 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:06 crc kubenswrapper[5109]: I1211 16:53:06.065270 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Dec 11 16:53:06 crc kubenswrapper[5109]: I1211 16:53:06.068497 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:06 crc kubenswrapper[5109]: I1211 16:53:06.069329 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:06 crc kubenswrapper[5109]: I1211 16:53:06.069391 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:06 crc kubenswrapper[5109]: I1211 16:53:06.069414 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:06 crc kubenswrapper[5109]: E1211 16:53:06.070056 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:06 crc kubenswrapper[5109]: I1211 16:53:06.070521 5109 scope.go:117] "RemoveContainer" containerID="8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d" Dec 11 16:53:06 crc kubenswrapper[5109]: E1211 16:53:06.070953 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:06 crc kubenswrapper[5109]: E1211 16:53:06.081179 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:06.070883206 +0000 UTC m=+23.750574712,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:06 crc kubenswrapper[5109]: I1211 16:53:06.815425 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:07 crc kubenswrapper[5109]: I1211 16:53:07.815576 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:08 crc kubenswrapper[5109]: I1211 16:53:08.815395 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:08 crc kubenswrapper[5109]: I1211 16:53:08.952433 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:08 crc kubenswrapper[5109]: I1211 16:53:08.953820 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:08 crc kubenswrapper[5109]: I1211 16:53:08.953896 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:08 crc kubenswrapper[5109]: I1211 16:53:08.953931 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:08 crc kubenswrapper[5109]: I1211 16:53:08.953984 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:53:08 crc kubenswrapper[5109]: E1211 16:53:08.969608 5109 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Dec 11 16:53:09 crc kubenswrapper[5109]: E1211 16:53:09.163946 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 11 16:53:09 crc kubenswrapper[5109]: E1211 16:53:09.455949 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 11 16:53:09 crc kubenswrapper[5109]: I1211 16:53:09.817190 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:10 crc kubenswrapper[5109]: I1211 16:53:10.816702 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:11 crc kubenswrapper[5109]: I1211 16:53:11.817098 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:12 crc kubenswrapper[5109]: E1211 16:53:12.455405 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 11 16:53:12 crc kubenswrapper[5109]: I1211 16:53:12.815387 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:12 crc kubenswrapper[5109]: E1211 16:53:12.971965 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:53:13 crc kubenswrapper[5109]: E1211 16:53:13.557169 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 11 16:53:13 crc kubenswrapper[5109]: I1211 16:53:13.815542 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.054616 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.056001 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.057271 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.057332 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.057354 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:14 crc kubenswrapper[5109]: E1211 16:53:14.057916 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.058284 5109 scope.go:117] "RemoveContainer" containerID="8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d" Dec 11 16:53:14 crc kubenswrapper[5109]: E1211 16:53:14.058577 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:14 crc kubenswrapper[5109]: E1211 16:53:14.066851 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:14.058531496 +0000 UTC m=+31.738222992,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.696954 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.697607 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.698940 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.699015 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.699046 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:14 crc kubenswrapper[5109]: E1211 16:53:14.699800 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.700270 5109 scope.go:117] "RemoveContainer" containerID="8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d" Dec 11 16:53:14 crc kubenswrapper[5109]: E1211 16:53:14.710330 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376635634f63\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376635634f63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.051127139 +0000 UTC m=+3.730818605,LastTimestamp:2025-12-11 16:53:14.701966432 +0000 UTC m=+32.381657938,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:14 crc kubenswrapper[5109]: I1211 16:53:14.814849 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:14 crc kubenswrapper[5109]: E1211 16:53:14.953182 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376643e494f3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376643e494f3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container: kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.294480115 +0000 UTC m=+3.974171591,LastTimestamp:2025-12-11 16:53:14.951778938 +0000 UTC m=+32.631470414,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:14 crc kubenswrapper[5109]: E1211 16:53:14.973513 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376644ca901a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376644ca901a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.309552154 +0000 UTC m=+3.989243630,LastTimestamp:2025-12-11 16:53:14.96724267 +0000 UTC m=+32.646934146,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.098446 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.101359 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04"} Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.101648 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.102981 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.103036 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.103057 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:15 crc kubenswrapper[5109]: E1211 16:53:15.103698 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:15 crc kubenswrapper[5109]: E1211 16:53:15.426447 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.815976 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.971052 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.972229 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.972303 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.972331 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:15 crc kubenswrapper[5109]: I1211 16:53:15.972379 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:53:15 crc kubenswrapper[5109]: E1211 16:53:15.985313 5109 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Dec 11 16:53:16 crc kubenswrapper[5109]: I1211 16:53:16.816655 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.109178 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.110045 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/1.log" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.113584 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04" exitCode=255 Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.113698 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04"} Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.113816 5109 scope.go:117] "RemoveContainer" containerID="8520ba0732981493baff1ee473ff392b68b4c4270eb02fd37741344953e5a12d" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.114145 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.115193 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.115260 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.115284 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:17 crc kubenswrapper[5109]: E1211 16:53:17.120256 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.120808 5109 scope.go:117] "RemoveContainer" containerID="5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04" Dec 11 16:53:17 crc kubenswrapper[5109]: E1211 16:53:17.121194 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:17 crc kubenswrapper[5109]: E1211 16:53:17.129300 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:17.121132954 +0000 UTC m=+34.800824460,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:17 crc kubenswrapper[5109]: I1211 16:53:17.815803 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:18 crc kubenswrapper[5109]: I1211 16:53:18.120026 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Dec 11 16:53:18 crc kubenswrapper[5109]: I1211 16:53:18.816379 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:19 crc kubenswrapper[5109]: E1211 16:53:19.464004 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 11 16:53:19 crc kubenswrapper[5109]: I1211 16:53:19.814873 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:20 crc kubenswrapper[5109]: I1211 16:53:20.817819 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:21 crc kubenswrapper[5109]: I1211 16:53:21.816674 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:22 crc kubenswrapper[5109]: I1211 16:53:22.814957 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:22 crc kubenswrapper[5109]: E1211 16:53:22.864680 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.RuntimeClass" Dec 11 16:53:22 crc kubenswrapper[5109]: E1211 16:53:22.973346 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:53:22 crc kubenswrapper[5109]: I1211 16:53:22.985870 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:22 crc kubenswrapper[5109]: I1211 16:53:22.986572 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:22 crc kubenswrapper[5109]: I1211 16:53:22.986615 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:22 crc kubenswrapper[5109]: I1211 16:53:22.986628 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:22 crc kubenswrapper[5109]: I1211 16:53:22.986658 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:53:23 crc kubenswrapper[5109]: E1211 16:53:23.001655 5109 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Dec 11 16:53:23 crc kubenswrapper[5109]: I1211 16:53:23.816585 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:24 crc kubenswrapper[5109]: I1211 16:53:24.696967 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:24 crc kubenswrapper[5109]: I1211 16:53:24.697274 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:24 crc kubenswrapper[5109]: I1211 16:53:24.698802 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:24 crc kubenswrapper[5109]: I1211 16:53:24.698870 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:24 crc kubenswrapper[5109]: I1211 16:53:24.698889 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:24 crc kubenswrapper[5109]: E1211 16:53:24.699440 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:24 crc kubenswrapper[5109]: I1211 16:53:24.699877 5109 scope.go:117] "RemoveContainer" containerID="5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04" Dec 11 16:53:24 crc kubenswrapper[5109]: E1211 16:53:24.700231 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:24 crc kubenswrapper[5109]: E1211 16:53:24.708627 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:24.700173697 +0000 UTC m=+42.379865203,Count:5,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:24 crc kubenswrapper[5109]: I1211 16:53:24.815588 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:25 crc kubenswrapper[5109]: I1211 16:53:25.102985 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:25 crc kubenswrapper[5109]: I1211 16:53:25.143883 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:25 crc kubenswrapper[5109]: I1211 16:53:25.144825 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:25 crc kubenswrapper[5109]: I1211 16:53:25.144885 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:25 crc kubenswrapper[5109]: I1211 16:53:25.144906 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:25 crc kubenswrapper[5109]: E1211 16:53:25.145479 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:25 crc kubenswrapper[5109]: I1211 16:53:25.146007 5109 scope.go:117] "RemoveContainer" containerID="5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04" Dec 11 16:53:25 crc kubenswrapper[5109]: E1211 16:53:25.146455 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:25 crc kubenswrapper[5109]: E1211 16:53:25.154794 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:25.146385818 +0000 UTC m=+42.826077324,Count:6,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:25 crc kubenswrapper[5109]: I1211 16:53:25.817360 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:26 crc kubenswrapper[5109]: E1211 16:53:26.472322 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 11 16:53:26 crc kubenswrapper[5109]: I1211 16:53:26.816781 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:27 crc kubenswrapper[5109]: I1211 16:53:27.815003 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:28 crc kubenswrapper[5109]: I1211 16:53:28.817534 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:29 crc kubenswrapper[5109]: I1211 16:53:29.814402 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.002377 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.003735 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.003835 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.003861 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.003905 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:53:30 crc kubenswrapper[5109]: E1211 16:53:30.018633 5109 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.381208 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.381535 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.382595 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.382662 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.382684 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:30 crc kubenswrapper[5109]: E1211 16:53:30.383197 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:30 crc kubenswrapper[5109]: I1211 16:53:30.813580 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:31 crc kubenswrapper[5109]: I1211 16:53:31.813644 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:32 crc kubenswrapper[5109]: I1211 16:53:32.813633 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:32 crc kubenswrapper[5109]: E1211 16:53:32.973862 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:53:33 crc kubenswrapper[5109]: E1211 16:53:33.350958 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Node" Dec 11 16:53:33 crc kubenswrapper[5109]: E1211 16:53:33.480208 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 11 16:53:33 crc kubenswrapper[5109]: I1211 16:53:33.814647 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:34 crc kubenswrapper[5109]: I1211 16:53:34.815640 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:34 crc kubenswrapper[5109]: E1211 16:53:34.876168 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.Service" Dec 11 16:53:35 crc kubenswrapper[5109]: I1211 16:53:35.816061 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:36 crc kubenswrapper[5109]: I1211 16:53:36.815886 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:36 crc kubenswrapper[5109]: I1211 16:53:36.900378 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:36 crc kubenswrapper[5109]: I1211 16:53:36.901945 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:36 crc kubenswrapper[5109]: I1211 16:53:36.902007 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:36 crc kubenswrapper[5109]: I1211 16:53:36.902021 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:36 crc kubenswrapper[5109]: E1211 16:53:36.902509 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:36 crc kubenswrapper[5109]: I1211 16:53:36.902950 5109 scope.go:117] "RemoveContainer" containerID="5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04" Dec 11 16:53:36 crc kubenswrapper[5109]: E1211 16:53:36.906479 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376635634f63\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376635634f63 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:52:46.051127139 +0000 UTC m=+3.730818605,LastTimestamp:2025-12-11 16:53:36.904785025 +0000 UTC m=+54.584476501,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.019084 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.021223 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.021294 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.021321 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.021360 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:53:37 crc kubenswrapper[5109]: E1211 16:53:37.031054 5109 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.178302 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.179994 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d"} Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.180203 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.180717 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.180804 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.180824 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:37 crc kubenswrapper[5109]: E1211 16:53:37.181362 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:37 crc kubenswrapper[5109]: I1211 16:53:37.815067 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:38 crc kubenswrapper[5109]: E1211 16:53:38.753963 5109 reflector.go:200] "Failed to watch" err="failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" reflector="k8s.io/client-go/informers/factory.go:160" type="*v1.CSIDriver" Dec 11 16:53:38 crc kubenswrapper[5109]: I1211 16:53:38.815265 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.187386 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.188275 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/2.log" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.190656 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" exitCode=255 Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.190722 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerDied","Data":"6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d"} Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.191177 5109 scope.go:117] "RemoveContainer" containerID="5e3ed2de65460a9024c274bc74389e6817e85c06ccc139903d5544d60ff3cf04" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.191492 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.192822 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.192869 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.192883 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:39 crc kubenswrapper[5109]: E1211 16:53:39.193281 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.193622 5109 scope.go:117] "RemoveContainer" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" Dec 11 16:53:39 crc kubenswrapper[5109]: E1211 16:53:39.194002 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:39 crc kubenswrapper[5109]: E1211 16:53:39.200277 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:39.193959466 +0000 UTC m=+56.873650952,Count:7,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:39 crc kubenswrapper[5109]: I1211 16:53:39.813149 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:40 crc kubenswrapper[5109]: I1211 16:53:40.195053 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Dec 11 16:53:40 crc kubenswrapper[5109]: E1211 16:53:40.487922 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 11 16:53:40 crc kubenswrapper[5109]: I1211 16:53:40.813207 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:41 crc kubenswrapper[5109]: I1211 16:53:41.817081 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:42 crc kubenswrapper[5109]: I1211 16:53:42.813580 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:42 crc kubenswrapper[5109]: E1211 16:53:42.974344 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:53:43 crc kubenswrapper[5109]: I1211 16:53:43.813697 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.031289 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.032523 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.032592 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.032632 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.032666 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:53:44 crc kubenswrapper[5109]: E1211 16:53:44.045386 5109 kubelet_node_status.go:116] "Unable to register node with API server, error getting existing node" err="nodes \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.696083 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.696287 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.697068 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.697399 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.697599 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:44 crc kubenswrapper[5109]: E1211 16:53:44.698459 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.699031 5109 scope.go:117] "RemoveContainer" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" Dec 11 16:53:44 crc kubenswrapper[5109]: E1211 16:53:44.699473 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:44 crc kubenswrapper[5109]: E1211 16:53:44.707587 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:44.699419223 +0000 UTC m=+62.379110719,Count:8,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:44 crc kubenswrapper[5109]: I1211 16:53:44.809530 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:45 crc kubenswrapper[5109]: I1211 16:53:45.814366 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:46 crc kubenswrapper[5109]: I1211 16:53:46.815800 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:47 crc kubenswrapper[5109]: I1211 16:53:47.180600 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:53:47 crc kubenswrapper[5109]: I1211 16:53:47.180871 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:47 crc kubenswrapper[5109]: I1211 16:53:47.181638 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:47 crc kubenswrapper[5109]: I1211 16:53:47.181678 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:47 crc kubenswrapper[5109]: I1211 16:53:47.181689 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:47 crc kubenswrapper[5109]: E1211 16:53:47.182321 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:47 crc kubenswrapper[5109]: I1211 16:53:47.182676 5109 scope.go:117] "RemoveContainer" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" Dec 11 16:53:47 crc kubenswrapper[5109]: E1211 16:53:47.184663 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:47 crc kubenswrapper[5109]: E1211 16:53:47.185382 5109 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.1880376aa28898fe\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.1880376aa28898fe openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:3a14caf222afb62aaabdc47808b6f944,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:BackOff,Message:Back-off restarting failed container kube-apiserver-check-endpoints in pod kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:53:05.062156542 +0000 UTC m=+22.741848018,LastTimestamp:2025-12-11 16:53:47.182951448 +0000 UTC m=+64.862642914,Count:9,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:53:47 crc kubenswrapper[5109]: E1211 16:53:47.496788 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Dec 11 16:53:47 crc kubenswrapper[5109]: I1211 16:53:47.812592 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:48 crc kubenswrapper[5109]: I1211 16:53:48.815276 5109 csi_plugin.go:988] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Dec 11 16:53:49 crc kubenswrapper[5109]: I1211 16:53:49.200073 5109 csr.go:274] "Certificate signing request is approved, waiting to be issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-w9qzw" Dec 11 16:53:49 crc kubenswrapper[5109]: I1211 16:53:49.207305 5109 csr.go:270] "Certificate signing request is issued" logger="kubernetes.io/kube-apiserver-client-kubelet" csr="csr-w9qzw" Dec 11 16:53:49 crc kubenswrapper[5109]: I1211 16:53:49.219313 5109 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 11 16:53:49 crc kubenswrapper[5109]: I1211 16:53:49.673728 5109 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Dec 11 16:53:50 crc kubenswrapper[5109]: I1211 16:53:50.208798 5109 certificate_manager.go:715] "Certificate rotation deadline determined" logger="kubernetes.io/kube-apiserver-client-kubelet" expiration="2026-01-10 16:48:49 +0000 UTC" deadline="2026-01-02 21:35:37.293690898 +0000 UTC" Dec 11 16:53:50 crc kubenswrapper[5109]: I1211 16:53:50.208903 5109 certificate_manager.go:431] "Waiting for next certificate rotation" logger="kubernetes.io/kube-apiserver-client-kubelet" sleep="532h41m47.084794052s" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.045978 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.047531 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.047864 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.048068 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.048440 5109 kubelet_node_status.go:78] "Attempting to register node" node="crc" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.060943 5109 kubelet_node_status.go:127] "Node was previously registered" node="crc" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.061419 5109 kubelet_node_status.go:81] "Successfully registered node" node="crc" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.061575 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.066429 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.066817 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.067044 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.067225 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.067369 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:53:51Z","lastTransitionTime":"2025-12-11T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.080762 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.088958 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.089026 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.089040 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.089065 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.089081 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:53:51Z","lastTransitionTime":"2025-12-11T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.107415 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.116480 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.116535 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.116551 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.116569 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.116582 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:53:51Z","lastTransitionTime":"2025-12-11T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.126676 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.135599 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.135646 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.135657 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.135671 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.135682 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:53:51Z","lastTransitionTime":"2025-12-11T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.146455 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.146588 5109 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.146638 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.246946 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.347687 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.448374 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.549016 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.650004 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.752272 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.852822 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.900192 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.901328 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.901443 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:51 crc kubenswrapper[5109]: I1211 16:53:51.901544 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.901946 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:51 crc kubenswrapper[5109]: E1211 16:53:51.953096 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.054165 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.155496 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.255661 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.356472 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.457449 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.558403 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.659157 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.759797 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.860842 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.961063 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:52 crc kubenswrapper[5109]: E1211 16:53:52.975367 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.061981 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.163164 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.263664 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.364680 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.465590 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.566515 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.666855 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.767470 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.868414 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:53 crc kubenswrapper[5109]: E1211 16:53:53.969346 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.070340 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.170850 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.271265 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.372608 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.472802 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.573565 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.674305 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.775021 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.875673 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:54 crc kubenswrapper[5109]: E1211 16:53:54.975838 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.076249 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.176820 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.277229 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.377810 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.478876 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.580046 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.681138 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.781426 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.882534 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:55 crc kubenswrapper[5109]: E1211 16:53:55.983568 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.084631 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.185697 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.286652 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.387363 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.487968 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.588333 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.688718 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.789135 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.890187 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:56 crc kubenswrapper[5109]: E1211 16:53:56.990939 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.091632 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.191971 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.292807 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.393530 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.493660 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.594587 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.695731 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.796634 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.897535 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:57 crc kubenswrapper[5109]: E1211 16:53:57.998151 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.099516 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.200602 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.301949 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.403809 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.504139 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.605373 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.706810 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.807378 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:58 crc kubenswrapper[5109]: E1211 16:53:58.908379 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.009577 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.110438 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.211138 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.311524 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.413033 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.513850 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.615084 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.715729 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.816517 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:53:59 crc kubenswrapper[5109]: I1211 16:53:59.899936 5109 kubelet_node_status.go:413] "Setting node annotation to enable volume controller attach/detach" Dec 11 16:53:59 crc kubenswrapper[5109]: I1211 16:53:59.901182 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:53:59 crc kubenswrapper[5109]: I1211 16:53:59.901321 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:53:59 crc kubenswrapper[5109]: I1211 16:53:59.901415 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.901936 5109 kubelet.go:3336] "No need to create a mirror pod, since failed to get node info from the cluster" err="node \"crc\" not found" node="crc" Dec 11 16:53:59 crc kubenswrapper[5109]: I1211 16:53:59.902279 5109 scope.go:117] "RemoveContainer" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.902561 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:53:59 crc kubenswrapper[5109]: E1211 16:53:59.916636 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.017703 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.118644 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.219338 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.320306 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.421862 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.522335 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.623614 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: I1211 16:54:00.722480 5109 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.723852 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.824851 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:00 crc kubenswrapper[5109]: E1211 16:54:00.925279 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.026216 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.127028 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.227264 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.327647 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.404232 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.411925 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.411982 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.412002 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.412023 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.412036 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:01Z","lastTransitionTime":"2025-12-11T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.466347 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.478878 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.478951 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.478971 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.479033 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.479055 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:01Z","lastTransitionTime":"2025-12-11T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.498541 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.503111 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.503158 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.503169 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.503187 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.503201 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:01Z","lastTransitionTime":"2025-12-11T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.514296 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.517941 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.517995 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.518007 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.518023 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:01 crc kubenswrapper[5109]: I1211 16:54:01.518033 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:01Z","lastTransitionTime":"2025-12-11T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.530770 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.530939 5109 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.530965 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.631628 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.731889 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.832443 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:01 crc kubenswrapper[5109]: E1211 16:54:01.933278 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.033786 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.134509 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.235209 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.336408 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.436962 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.538258 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.639151 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.739539 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.839796 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.939914 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:02 crc kubenswrapper[5109]: E1211 16:54:02.976307 5109 eviction_manager.go:292] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.040453 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.141282 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.242023 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.342641 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.443103 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.543611 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.644292 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.745026 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.845354 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:03 crc kubenswrapper[5109]: E1211 16:54:03.945509 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.046102 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.146841 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.247338 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.348367 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.449402 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.550451 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.651098 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.752234 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.852568 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:04 crc kubenswrapper[5109]: E1211 16:54:04.952789 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.053053 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.153669 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.253865 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.355258 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.455841 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.557005 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.657406 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.758433 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.859200 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:05 crc kubenswrapper[5109]: E1211 16:54:05.960071 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.061105 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.161924 5109 kubelet_node_status.go:515] "Error getting the current node from lister" err="node \"crc\" not found" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.248515 5109 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.264448 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.264512 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.264534 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.264560 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.264579 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.325184 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.340056 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.368207 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.368275 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.368290 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.368312 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.368325 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.440892 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.471399 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.471464 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.471487 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.471555 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.471573 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.542946 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.574007 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.574069 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.574087 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.574110 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.574130 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.644941 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-etcd/etcd-crc" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.677158 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.677206 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.677218 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.677235 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.677250 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.780106 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.780160 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.780173 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.780195 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.780209 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.832325 5109 apiserver.go:52] "Watching apiserver" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.846513 5109 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.849008 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc","openshift-image-registry/node-ca-xqdz5","openshift-kube-apiserver/kube-apiserver-crc","openshift-network-node-identity/network-node-identity-dgvkt","openshift-network-operator/iptables-alerter-5jnd7","openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv","openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n","openshift-dns/node-resolver-rnsbf","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-multus/multus-hx2kb","openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6","openshift-machine-config-operator/machine-config-daemon-xgkvv","openshift-multus/multus-additional-cni-plugins-c2wzl","openshift-multus/network-metrics-daemon-fp2cw","openshift-network-diagnostics/network-check-target-fhkjl","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5","openshift-ovn-kubernetes/ovnkube-node-wfxmd"] Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.851630 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.852701 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.852863 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.854096 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.854167 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.854532 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.854833 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.855690 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"metrics-tls\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.856387 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.857789 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.858677 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"env-overrides\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.858894 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-node-identity\"/\"network-node-identity-cert\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.859089 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"ovnkube-identity-cm\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.859350 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"openshift-service-ca.crt\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.859634 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.859792 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.860183 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"kube-root-ca.crt\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.860428 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.872278 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.885239 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.885361 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.885377 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.885396 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.885407 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.889582 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.903303 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.922951 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.931436 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.931605 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1f09ed1-77b5-404d-abd1-44afa5fab784-proxy-tls\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.931710 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1f09ed1-77b5-404d-abd1-44afa5fab784-mcd-auth-proxy-config\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.931856 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.931962 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.932095 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c1f09ed1-77b5-404d-abd1-44afa5fab784-rootfs\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.932217 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.932341 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vmt7\" (UniqueName: \"kubernetes.io/projected/c1f09ed1-77b5-404d-abd1-44afa5fab784-kube-api-access-5vmt7\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.932468 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.932590 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.932728 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.932917 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-8nt2j\" (UniqueName: \"kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.933024 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.933126 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.933222 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-m7xz2\" (UniqueName: \"kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.933324 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.933417 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dsgwk\" (UniqueName: \"kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.933516 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.933711 5109 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.934652 5109 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.934931 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.433889717 +0000 UTC m=+85.113581203 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.935178 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.435155808 +0000 UTC m=+85.114847284 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.947399 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.953203 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.953241 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.953256 5109 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.953348 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.453326056 +0000 UTC m=+85.133017532 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.962657 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.962687 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.962700 5109 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:06 crc kubenswrapper[5109]: E1211 16:54:06.962791 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.462772324 +0000 UTC m=+85.142463800 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.966350 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.978664 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.987181 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.987226 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.987239 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.987261 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:06 crc kubenswrapper[5109]: I1211 16:54:06.987282 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:06Z","lastTransitionTime":"2025-12-11T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034365 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1f09ed1-77b5-404d-abd1-44afa5fab784-proxy-tls\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034432 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1f09ed1-77b5-404d-abd1-44afa5fab784-mcd-auth-proxy-config\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034478 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034515 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c1f09ed1-77b5-404d-abd1-44afa5fab784-rootfs\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034554 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.034580 5109 secret.go:189] Couldn't get secret openshift-machine-config-operator/proxy-tls: object "openshift-machine-config-operator"/"proxy-tls" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.034595 5109 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: object "openshift-machine-config-operator"/"kube-rbac-proxy" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.034678 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c1f09ed1-77b5-404d-abd1-44afa5fab784-proxy-tls podName:c1f09ed1-77b5-404d-abd1-44afa5fab784 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.534656249 +0000 UTC m=+85.214347715 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/c1f09ed1-77b5-404d-abd1-44afa5fab784-proxy-tls") pod "machine-config-daemon-xgkvv" (UID: "c1f09ed1-77b5-404d-abd1-44afa5fab784") : object "openshift-machine-config-operator"/"proxy-tls" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.034702 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c1f09ed1-77b5-404d-abd1-44afa5fab784-mcd-auth-proxy-config podName:c1f09ed1-77b5-404d-abd1-44afa5fab784 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.53469186 +0000 UTC m=+85.214383326 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "mcd-auth-proxy-config" (UniqueName: "kubernetes.io/configmap/c1f09ed1-77b5-404d-abd1-44afa5fab784-mcd-auth-proxy-config") pod "machine-config-daemon-xgkvv" (UID: "c1f09ed1-77b5-404d-abd1-44afa5fab784") : object "openshift-machine-config-operator"/"kube-rbac-proxy" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034585 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5vmt7\" (UniqueName: \"kubernetes.io/projected/c1f09ed1-77b5-404d-abd1-44afa5fab784-kube-api-access-5vmt7\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034785 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/c1f09ed1-77b5-404d-abd1-44afa5fab784-rootfs\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034893 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-host-slash\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.034942 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/34177974-8d82-49d2-a763-391d0df3bbd8-host-etc-kube\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.059470 5109 projected.go:289] Couldn't get configMap openshift-machine-config-operator/kube-root-ca.crt: object "openshift-machine-config-operator"/"kube-root-ca.crt" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.059556 5109 projected.go:289] Couldn't get configMap openshift-machine-config-operator/openshift-service-ca.crt: object "openshift-machine-config-operator"/"openshift-service-ca.crt" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.059581 5109 projected.go:194] Error preparing data for projected volume kube-api-access-5vmt7 for pod openshift-machine-config-operator/machine-config-daemon-xgkvv: [object "openshift-machine-config-operator"/"kube-root-ca.crt" not registered, object "openshift-machine-config-operator"/"openshift-service-ca.crt" not registered] Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.059698 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c1f09ed1-77b5-404d-abd1-44afa5fab784-kube-api-access-5vmt7 podName:c1f09ed1-77b5-404d-abd1-44afa5fab784 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.559663573 +0000 UTC m=+85.239355079 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-5vmt7" (UniqueName: "kubernetes.io/projected/c1f09ed1-77b5-404d-abd1-44afa5fab784-kube-api-access-5vmt7") pod "machine-config-daemon-xgkvv" (UID: "c1f09ed1-77b5-404d-abd1-44afa5fab784") : [object "openshift-machine-config-operator"/"kube-root-ca.crt" not registered, object "openshift-machine-config-operator"/"openshift-service-ca.crt" not registered] Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.088613 5109 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.089397 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-env-overrides\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.090370 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.090417 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.090430 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.090449 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.090465 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.093665 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-iptables-alerter-script\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.093998 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/fc4541ce-7789-4670-bc75-5c2868e52ce0-ovnkube-identity-cm\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.098543 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7xz2\" (UniqueName: \"kubernetes.io/projected/34177974-8d82-49d2-a763-391d0df3bbd8-kube-api-access-m7xz2\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.098546 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/34177974-8d82-49d2-a763-391d0df3bbd8-metrics-tls\") pod \"network-operator-7bdcf4f5bd-7fjxv\" (UID: \"34177974-8d82-49d2-a763-391d0df3bbd8\") " pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.099196 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nt2j\" (UniqueName: \"kubernetes.io/projected/fc4541ce-7789-4670-bc75-5c2868e52ce0-kube-api-access-8nt2j\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.099473 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dsgwk\" (UniqueName: \"kubernetes.io/projected/428b39f5-eb1c-4f65-b7a4-eeb6e84860cc-kube-api-access-dsgwk\") pod \"iptables-alerter-5jnd7\" (UID: \"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\") " pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.100238 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fc4541ce-7789-4670-bc75-5c2868e52ce0-webhook-cert\") pod \"network-node-identity-dgvkt\" (UID: \"fc4541ce-7789-4670-bc75-5c2868e52ce0\") " pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.100631 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.104041 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"proxy-tls\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.104243 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.104701 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.105213 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-w9nzh\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.105819 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.112874 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.114405 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.115222 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.116177 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-tk7bt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.121594 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.124278 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.124362 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.124606 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.124876 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.125420 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-g6kgg\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.126664 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.130254 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-nwglk\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.130291 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.130392 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.130433 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.131457 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.131649 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135112 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135200 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-os-release\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135324 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-cni-binary-copy\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135388 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6lpl\" (UniqueName: \"kubernetes.io/projected/189d054c-6301-4cf4-ba60-da0380233d01-kube-api-access-l6lpl\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135427 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-k8s-cni-cncf-io\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135464 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-cnibin\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135500 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-cni-bin\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135566 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-tmp-dir\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135684 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbrkp\" (UniqueName: \"kubernetes.io/projected/7c45b4ae-02aa-4614-977b-544a9e303bdf-kube-api-access-nbrkp\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.135831 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-system-cni-dir\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137531 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-cni-multus\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137588 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-system-cni-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137620 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-daemon-config\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137700 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-multus-certs\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137761 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-cnibin\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137796 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-os-release\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137838 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-socket-dir-parent\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137884 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-kubelet\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137937 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6v94n\" (UniqueName: \"kubernetes.io/projected/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-kube-api-access-6v94n\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.137967 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-etc-kubernetes\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138019 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138072 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-hosts-file\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138100 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7c45b4ae-02aa-4614-977b-544a9e303bdf-cni-binary-copy\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138132 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-netns\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138181 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138210 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-hostroot\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138280 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-cni-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.138308 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-conf-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.139501 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.142181 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.142379 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-l2v2m\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.142458 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.142945 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.142931 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.145600 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.148625 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.155246 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.155666 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.155698 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.157096 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-nl8tp\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.165835 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.172816 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.177931 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1f09ed1-77b5-404d-abd1-44afa5fab784\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xgkvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.181166 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-dgvkt" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.194259 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.194339 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.194362 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.194393 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.194415 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.197687 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.201894 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,Command:[/bin/bash -c #!/bin/bash Dec 11 16:54:07 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: source /etc/kubernetes/apiserver-url.env Dec 11 16:54:07 crc kubenswrapper[5109]: else Dec 11 16:54:07 crc kubenswrapper[5109]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Dec 11 16:54:07 crc kubenswrapper[5109]: exit 1 Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.20.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:951276a60f15185a05902cf1ec49b6db3e4f049ec638828b336aed496f8dfc45,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b5000f8f055fd8f734ef74afbd9bd5333a38345cbc4959ddaad728b8394bccd4,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be136d591a0eeb3f7bedf04aabb5481a23b6645316d5cef3cd5be1787344c2b5,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91997a073272252cac9cd31915ec74217637c55d1abc725107c6eb677ddddc9b,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6a974f04d4aefdb39bf2d4649b24e7e0e87685afa3d07ca46234f1a0c5688e4b,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m7xz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-7bdcf4f5bd-7fjxv_openshift-network-operator(34177974-8d82-49d2-a763-391d0df3bbd8): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: W1211 16:54:07.201905 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfc4541ce_7789_4670_bc75_5c2868e52ce0.slice/crio-ad647ed47fee506c3ae19e763c92a6f718b792d99dc8b74f3d37af7e5190c95b WatchSource:0}: Error finding container ad647ed47fee506c3ae19e763c92a6f718b792d99dc8b74f3d37af7e5190c95b: Status 404 returned error can't find the container with id ad647ed47fee506c3ae19e763c92a6f718b792d99dc8b74f3d37af7e5190c95b Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.203044 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" podUID="34177974-8d82-49d2-a763-391d0df3bbd8" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.208038 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c set -xe Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -f "/env/_master" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: source "/env/_master" Dec 11 16:54:07 crc kubenswrapper[5109]: set +o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Dec 11 16:54:07 crc kubenswrapper[5109]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Dec 11 16:54:07 crc kubenswrapper[5109]: ho_enable="--enable-hybrid-overlay" Dec 11 16:54:07 crc kubenswrapper[5109]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Dec 11 16:54:07 crc kubenswrapper[5109]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Dec 11 16:54:07 crc kubenswrapper[5109]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --webhook-cert-dir="/etc/webhook-cert" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --webhook-host=127.0.0.1 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --webhook-port=9743 \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${ho_enable} \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-interconnect \ Dec 11 16:54:07 crc kubenswrapper[5109]: --disable-approver \ Dec 11 16:54:07 crc kubenswrapper[5109]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --wait-for-kubernetes-api=200s \ Dec 11 16:54:07 crc kubenswrapper[5109]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --loglevel="${LOGLEVEL}" Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nt2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000500000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-dgvkt_openshift-network-node-identity(fc4541ce-7789-4670-bc75-5c2868e52ce0): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.211414 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c set -xe Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -f "/env/_master" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: source "/env/_master" Dec 11 16:54:07 crc kubenswrapper[5109]: set +o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --disable-webhook \ Dec 11 16:54:07 crc kubenswrapper[5109]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --loglevel="${LOGLEVEL}" Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nt2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000500000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-dgvkt_openshift-network-node-identity(fc4541ce-7789-4670-bc75-5c2868e52ce0): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.212566 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-dgvkt" podUID="fc4541ce-7789-4670-bc75-5c2868e52ce0" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.212555 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.225382 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.236955 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1f09ed1-77b5-404d-abd1-44afa5fab784\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xgkvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239196 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-multus-certs\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239264 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-var-lib-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239293 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-multus-certs\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239301 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61fc568b-d7c2-4633-b28c-f65428d01030-host\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239374 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4fpt\" (UniqueName: \"kubernetes.io/projected/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-kube-api-access-c4fpt\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239450 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-cnibin\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239491 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/561da642-b358-4ddc-86ee-669164d7b3c5-ovn-node-metrics-cert\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239525 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-node-log\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239537 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-cnibin\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239588 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-ovn-kubernetes\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239614 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/61fc568b-d7c2-4633-b28c-f65428d01030-serviceca\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239640 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-os-release\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239657 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-socket-dir-parent\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239674 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-kubelet\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239727 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-kubelet\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239784 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-os-release\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239800 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6v94n\" (UniqueName: \"kubernetes.io/projected/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-kube-api-access-6v94n\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239851 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-etc-kubernetes\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239857 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-socket-dir-parent\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239890 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-kubelet\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239924 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-env-overrides\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.239967 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-etc-kubernetes\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240057 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240129 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-hosts-file\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240226 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7c45b4ae-02aa-4614-977b-544a9e303bdf-cni-binary-copy\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240227 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-hosts-file\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240271 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-netns\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240311 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-netns\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240331 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-systemd\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240371 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240407 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-hostroot\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240440 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-netns\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240472 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240501 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-hostroot\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240509 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240640 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-whereabouts-flatfile-configmap\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240708 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-cni-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240672 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-cni-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240813 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-conf-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240864 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-systemd-units\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240931 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-conf-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240955 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-bin\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.240960 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7c45b4ae-02aa-4614-977b-544a9e303bdf-cni-binary-copy\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241030 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241233 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-os-release\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241239 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241286 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-os-release\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241286 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241373 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-slash\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241411 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-netd\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241599 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-script-lib\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241643 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241685 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-cni-binary-copy\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241720 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l6lpl\" (UniqueName: \"kubernetes.io/projected/189d054c-6301-4cf4-ba60-da0380233d01-kube-api-access-l6lpl\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241796 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-k8s-cni-cncf-io\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241922 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-run-k8s-cni-cncf-io\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241844 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.241995 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-cnibin\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242195 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-cnibin\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242295 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-cni-bin\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242340 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242376 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-config\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242432 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-tmp-dir\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242465 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-log-socket\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242495 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46pxq\" (UniqueName: \"kubernetes.io/projected/561da642-b358-4ddc-86ee-669164d7b3c5-kube-api-access-46pxq\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242537 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nbrkp\" (UniqueName: \"kubernetes.io/projected/7c45b4ae-02aa-4614-977b-544a9e303bdf-kube-api-access-nbrkp\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242570 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-etc-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242599 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p72wg\" (UniqueName: \"kubernetes.io/projected/61fc568b-d7c2-4633-b28c-f65428d01030-kube-api-access-p72wg\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242636 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-system-cni-dir\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242675 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-cni-multus\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242705 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-ovn\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242775 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242817 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-system-cni-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242852 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-daemon-config\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242886 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5rg8\" (UniqueName: \"kubernetes.io/projected/ad4ed271-a1a8-408d-b043-99cf044fb573-kube-api-access-x5rg8\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.242953 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-cni-bin\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.243446 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/189d054c-6301-4cf4-ba60-da0380233d01-cni-binary-copy\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.243622 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-tmp-dir\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.243734 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/189d054c-6301-4cf4-ba60-da0380233d01-system-cni-dir\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.243859 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-host-var-lib-cni-multus\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.243986 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/7c45b4ae-02aa-4614-977b-544a9e303bdf-system-cni-dir\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.246369 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/7c45b4ae-02aa-4614-977b-544a9e303bdf-multus-daemon-config\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.251353 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.252461 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.253952 5109 scope.go:117] "RemoveContainer" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.254212 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-tjs74\"" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.254253 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.255192 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.256556 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.256565 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.258061 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6v94n\" (UniqueName: \"kubernetes.io/projected/305d0e9b-83ad-424c-a5a5-c4ba12d3ae45-kube-api-access-6v94n\") pod \"node-resolver-rnsbf\" (UID: \"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\") " pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.264920 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.265514 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" event={"ID":"34177974-8d82-49d2-a763-391d0df3bbd8","Type":"ContainerStarted","Data":"9277b66f35d63ac61842c5a6b88152b555d38ba1d0912b6a5ba04a9b4243f32f"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.266508 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"ad647ed47fee506c3ae19e763c92a6f718b792d99dc8b74f3d37af7e5190c95b"} Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.268609 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c set -xe Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -f "/env/_master" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: source "/env/_master" Dec 11 16:54:07 crc kubenswrapper[5109]: set +o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Dec 11 16:54:07 crc kubenswrapper[5109]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Dec 11 16:54:07 crc kubenswrapper[5109]: ho_enable="--enable-hybrid-overlay" Dec 11 16:54:07 crc kubenswrapper[5109]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Dec 11 16:54:07 crc kubenswrapper[5109]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Dec 11 16:54:07 crc kubenswrapper[5109]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --webhook-cert-dir="/etc/webhook-cert" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --webhook-host=127.0.0.1 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --webhook-port=9743 \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${ho_enable} \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-interconnect \ Dec 11 16:54:07 crc kubenswrapper[5109]: --disable-approver \ Dec 11 16:54:07 crc kubenswrapper[5109]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --wait-for-kubernetes-api=200s \ Dec 11 16:54:07 crc kubenswrapper[5109]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --loglevel="${LOGLEVEL}" Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nt2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000500000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-dgvkt_openshift-network-node-identity(fc4541ce-7789-4670-bc75-5c2868e52ce0): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.268760 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,Command:[/bin/bash -c #!/bin/bash Dec 11 16:54:07 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: source /etc/kubernetes/apiserver-url.env Dec 11 16:54:07 crc kubenswrapper[5109]: else Dec 11 16:54:07 crc kubenswrapper[5109]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Dec 11 16:54:07 crc kubenswrapper[5109]: exit 1 Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.20.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:951276a60f15185a05902cf1ec49b6db3e4f049ec638828b336aed496f8dfc45,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b5000f8f055fd8f734ef74afbd9bd5333a38345cbc4959ddaad728b8394bccd4,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be136d591a0eeb3f7bedf04aabb5481a23b6645316d5cef3cd5be1787344c2b5,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91997a073272252cac9cd31915ec74217637c55d1abc725107c6eb677ddddc9b,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6a974f04d4aefdb39bf2d4649b24e7e0e87685afa3d07ca46234f1a0c5688e4b,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-m7xz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-7bdcf4f5bd-7fjxv_openshift-network-operator(34177974-8d82-49d2-a763-391d0df3bbd8): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.270448 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" podUID="34177974-8d82-49d2-a763-391d0df3bbd8" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.271893 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c set -xe Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -f "/env/_master" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: source "/env/_master" Dec 11 16:54:07 crc kubenswrapper[5109]: set +o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --disable-webhook \ Dec 11 16:54:07 crc kubenswrapper[5109]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --loglevel="${LOGLEVEL}" Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8nt2j,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000500000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-dgvkt_openshift-network-node-identity(fc4541ce-7789-4670-bc75-5c2868e52ce0): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.272130 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6lpl\" (UniqueName: \"kubernetes.io/projected/189d054c-6301-4cf4-ba60-da0380233d01-kube-api-access-l6lpl\") pod \"multus-additional-cni-plugins-c2wzl\" (UID: \"189d054c-6301-4cf4-ba60-da0380233d01\") " pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.273303 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-dgvkt" podUID="fc4541ce-7789-4670-bc75-5c2868e52ce0" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.274196 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbrkp\" (UniqueName: \"kubernetes.io/projected/7c45b4ae-02aa-4614-977b-544a9e303bdf-kube-api-access-nbrkp\") pod \"multus-hx2kb\" (UID: \"7c45b4ae-02aa-4614-977b-544a9e303bdf\") " pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.279107 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-hx2kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c45b4ae-02aa-4614-977b-544a9e303bdf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbrkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hx2kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.296540 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.296615 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.296628 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.296673 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.296687 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.297557 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561da642-b358-4ddc-86ee-669164d7b3c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wfxmd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.309578 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-dzv7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.322908 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.326385 5109 desired_state_of_world_populator.go:158] "Finished populating initial desired state of world" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.334205 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.343888 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.343981 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344006 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344035 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4smf\" (UniqueName: \"kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf\") pod \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\" (UID: \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344058 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344082 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344104 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344138 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344168 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftwb6\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344193 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344218 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dmhf\" (UniqueName: \"kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344241 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344265 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-94l9h\" (UniqueName: \"kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344285 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344310 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wj4qr\" (UniqueName: \"kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344331 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344353 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344375 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344396 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344419 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344465 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344489 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hm9x7\" (UniqueName: \"kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344509 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.344531 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345497 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7" (OuterVolumeSpecName: "kube-api-access-hm9x7") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "kube-api-access-hm9x7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345556 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345622 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345819 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf" (OuterVolumeSpecName: "kube-api-access-6dmhf") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "kube-api-access-6dmhf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345866 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345861 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345921 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345949 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345996 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346019 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346062 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346085 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346107 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfp5s\" (UniqueName: \"kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346154 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346178 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346232 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346257 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346281 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346573 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grwfz\" (UniqueName: \"kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346630 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346670 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbmqg\" (UniqueName: \"kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346713 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346786 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346827 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346866 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346906 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346943 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346978 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pllx6\" (UniqueName: \"kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347019 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert\") pod \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\" (UID: \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347054 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347089 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content\") pod \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\" (UID: \"31fa8943-81cc-4750-a0b7-0fa9ab5af883\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347128 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls\") pod \"42a11a02-47e1-488f-b270-2679d3298b0e\" (UID: \"42a11a02-47e1-488f-b270-2679d3298b0e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347182 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347240 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347325 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqbfk\" (UniqueName: \"kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348272 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348369 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348434 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tknt7\" (UniqueName: \"kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348486 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348961 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities\") pod \"cc85e424-18b2-4924-920b-bd291a8c4b01\" (UID: \"cc85e424-18b2-4924-920b-bd291a8c4b01\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349031 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349077 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349114 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349153 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349195 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349228 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349259 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349315 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9z4sw\" (UniqueName: \"kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349346 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6rmnv\" (UniqueName: \"kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349380 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349565 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349603 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349630 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsb9b\" (UniqueName: \"kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349664 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349698 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349727 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmmzf\" (UniqueName: \"kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349840 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349876 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349921 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349963 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349999 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jjkz\" (UniqueName: \"kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350034 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350063 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350097 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350133 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350168 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350195 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350225 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g8ts\" (UniqueName: \"kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350255 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs\") pod \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\" (UID: \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350283 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350316 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350358 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzt4w\" (UniqueName: \"kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w\") pod \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\" (UID: \"a52afe44-fb37-46ed-a1f8-bf39727a3cbe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345907 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.345919 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346194 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr" (OuterVolumeSpecName: "kube-api-access-wj4qr") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "kube-api-access-wj4qr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346367 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h" (OuterVolumeSpecName: "kube-api-access-94l9h") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "kube-api-access-94l9h". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346458 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346493 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config" (OuterVolumeSpecName: "config") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350707 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6" (OuterVolumeSpecName: "kube-api-access-pllx6") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "kube-api-access-pllx6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346640 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.346803 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347320 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347885 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config" (OuterVolumeSpecName: "config") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348143 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348231 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348244 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf" (OuterVolumeSpecName: "kube-api-access-q4smf") pod "0dd0fbac-8c0d-4228-8faa-abbeedabf7db" (UID: "0dd0fbac-8c0d-4228-8faa-abbeedabf7db"). InnerVolumeSpecName "kube-api-access-q4smf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348414 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347349 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350826 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6" (OuterVolumeSpecName: "kube-api-access-ftwb6") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "kube-api-access-ftwb6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347348 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350919 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348633 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp" (OuterVolumeSpecName: "tmp") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.348649 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config" (OuterVolumeSpecName: "config") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349111 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349284 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349362 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349371 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349873 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7" (OuterVolumeSpecName: "kube-api-access-tknt7") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "kube-api-access-tknt7". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.349906 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s" (OuterVolumeSpecName: "kube-api-access-xfp5s") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "kube-api-access-xfp5s". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350101 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350457 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350462 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347988 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config" (OuterVolumeSpecName: "config") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.350964 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351168 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351282 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351506 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351469 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "42a11a02-47e1-488f-b270-2679d3298b0e" (UID: "42a11a02-47e1-488f-b270-2679d3298b0e"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351707 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351725 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities" (OuterVolumeSpecName: "utilities") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351963 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351975 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz" (OuterVolumeSpecName: "kube-api-access-grwfz") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "kube-api-access-grwfz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.351860 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities" (OuterVolumeSpecName: "utilities") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.352159 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.352055 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.352392 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert" (OuterVolumeSpecName: "cert") pod "a52afe44-fb37-46ed-a1f8-bf39727a3cbe" (UID: "a52afe44-fb37-46ed-a1f8-bf39727a3cbe"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.352395 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w" (OuterVolumeSpecName: "kube-api-access-rzt4w") pod "a52afe44-fb37-46ed-a1f8-bf39727a3cbe" (UID: "a52afe44-fb37-46ed-a1f8-bf39727a3cbe"). InnerVolumeSpecName "kube-api-access-rzt4w". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.352715 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.352821 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.352857 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.353124 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.353194 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.353248 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv" (OuterVolumeSpecName: "kube-api-access-6rmnv") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "kube-api-access-6rmnv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.353330 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.353684 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume" (OuterVolumeSpecName: "config-volume") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.353907 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354076 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354217 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354308 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg" (OuterVolumeSpecName: "kube-api-access-wbmqg") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "kube-api-access-wbmqg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354335 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354441 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities" (OuterVolumeSpecName: "utilities") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354678 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zth6t\" (UniqueName: \"kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t\") pod \"6077b63e-53a2-4f96-9d56-1ce0324e4913\" (UID: \"6077b63e-53a2-4f96-9d56-1ce0324e4913\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354810 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.347304 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.357459 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.357526 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354681 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.354873 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw" (OuterVolumeSpecName: "kube-api-access-9z4sw") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "kube-api-access-9z4sw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.355160 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.355561 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp" (OuterVolumeSpecName: "tmp") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.355678 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.355905 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config" (OuterVolumeSpecName: "config") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356020 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356133 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356160 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk" (OuterVolumeSpecName: "kube-api-access-qqbfk") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "kube-api-access-qqbfk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.358210 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356335 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz" (OuterVolumeSpecName: "kube-api-access-7jjkz") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "kube-api-access-7jjkz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356339 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356680 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config" (OuterVolumeSpecName: "config") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356772 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b" (OuterVolumeSpecName: "kube-api-access-zsb9b") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "kube-api-access-zsb9b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.356987 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.357556 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.358451 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.358483 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.357448 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts" (OuterVolumeSpecName: "kube-api-access-4g8ts") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "kube-api-access-4g8ts". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.358550 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.358550 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" (UID: "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.358615 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf" (OuterVolumeSpecName: "kube-api-access-nmmzf") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "kube-api-access-nmmzf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.359102 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.359005 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxfcv\" (UniqueName: \"kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.359788 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t" (OuterVolumeSpecName: "kube-api-access-zth6t") pod "6077b63e-53a2-4f96-9d56-1ce0324e4913" (UID: "6077b63e-53a2-4f96-9d56-1ce0324e4913"). InnerVolumeSpecName "kube-api-access-zth6t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.359857 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config" (OuterVolumeSpecName: "console-config") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.359977 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.359975 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.360634 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities" (OuterVolumeSpecName: "utilities") pod "cc85e424-18b2-4924-920b-bd291a8c4b01" (UID: "cc85e424-18b2-4924-920b-bd291a8c4b01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.360867 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.360900 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.360879 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap" (OuterVolumeSpecName: "whereabouts-flatfile-configmap") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "whereabouts-flatfile-configmap". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.361208 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca" (OuterVolumeSpecName: "client-ca") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.361326 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv" (OuterVolumeSpecName: "kube-api-access-xxfcv") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "kube-api-access-xxfcv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.361539 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs" (OuterVolumeSpecName: "certs") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.361659 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.361795 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.361811 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config" (OuterVolumeSpecName: "config") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.362986 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363029 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363055 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363078 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hckvg\" (UniqueName: \"kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363102 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363125 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pddnv\" (UniqueName: \"kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv\") pod \"e093be35-bb62-4843-b2e8-094545761610\" (UID: \"e093be35-bb62-4843-b2e8-094545761610\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363149 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkdh6\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363172 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates\") pod \"16bdd140-dce1-464c-ab47-dd5798d1d256\" (UID: \"16bdd140-dce1-464c-ab47-dd5798d1d256\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363195 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities\") pod \"149b3c48-e17c-4a66-a835-d86dabf6ff13\" (UID: \"149b3c48-e17c-4a66-a835-d86dabf6ff13\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.363220 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides\") pod \"7df94c10-441d-4386-93a6-6730fb7bcde0\" (UID: \"7df94c10-441d-4386-93a6-6730fb7bcde0\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.364090 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.364402 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dztfv\" (UniqueName: \"kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv\") pod \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\" (UID: \"d45be74c-0d98-4d18-90e4-f7ef1b6daaf7\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.364448 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.364520 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg" (OuterVolumeSpecName: "kube-api-access-hckvg") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "kube-api-access-hckvg". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.364827 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365201 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365234 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365541 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config" (OuterVolumeSpecName: "config") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365614 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365788 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365829 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv" (OuterVolumeSpecName: "kube-api-access-pddnv") pod "e093be35-bb62-4843-b2e8-094545761610" (UID: "e093be35-bb62-4843-b2e8-094545761610"). InnerVolumeSpecName "kube-api-access-pddnv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365847 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365906 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365935 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.365992 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366019 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366078 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366105 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ws8zz\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366163 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366191 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366245 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjwtd\" (UniqueName: \"kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366275 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twvbl\" (UniqueName: \"kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366298 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366361 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptkcf\" (UniqueName: \"kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf\") pod \"7599e0b6-bddf-4def-b7f2-0b32206e8651\" (UID: \"7599e0b6-bddf-4def-b7f2-0b32206e8651\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366387 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir\") pod \"f7e2c886-118e-43bb-bef1-c78134de392b\" (UID: \"f7e2c886-118e-43bb-bef1-c78134de392b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366430 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366491 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366519 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities\") pod \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\" (UID: \"9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366541 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366588 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g4lr\" (UniqueName: \"kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr\") pod \"f7e2c886-118e-43bb-bef1-c78134de392b\" (UID: \"f7e2c886-118e-43bb-bef1-c78134de392b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366610 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs\") pod \"18f80adb-c1c3-49ba-8ee4-932c851d3897\" (UID: \"18f80adb-c1c3-49ba-8ee4-932c851d3897\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366666 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366689 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs\") pod \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\" (UID: \"0dd0fbac-8c0d-4228-8faa-abbeedabf7db\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366748 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7cps\" (UniqueName: \"kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps\") pod \"af41de71-79cf-4590-bbe9-9e8b848862cb\" (UID: \"af41de71-79cf-4590-bbe9-9e8b848862cb\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366774 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366820 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366846 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366869 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nb9c\" (UniqueName: \"kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366913 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert\") pod \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\" (UID: \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366940 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366993 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnxbn\" (UniqueName: \"kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367052 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4hb7m\" (UniqueName: \"kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367078 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367217 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367262 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367309 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp\") pod \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\" (UID: \"b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367334 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgx6b\" (UniqueName: \"kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b\") pod \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\" (UID: \"f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367381 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content\") pod \"94a6e063-3d1a-4d44-875d-185291448c31\" (UID: \"94a6e063-3d1a-4d44-875d-185291448c31\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367408 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367456 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pskd\" (UniqueName: \"kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd\") pod \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\" (UID: \"a555ff2e-0be6-46d5-897d-863bb92ae2b3\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367481 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367505 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbc2l\" (UniqueName: \"kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367551 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l87hs\" (UniqueName: \"kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs\") pod \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\" (UID: \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367578 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367626 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26xrl\" (UniqueName: \"kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl\") pod \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\" (UID: \"a208c9c2-333b-4b4a-be0d-bc32ec38a821\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367654 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367774 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config\") pod \"81e39f7b-62e4-4fc9-992a-6535ce127a02\" (UID: \"81e39f7b-62e4-4fc9-992a-6535ce127a02\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367803 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367864 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367892 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368111 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368167 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp\") pod \"7afa918d-be67-40a6-803c-d3b0ae99d815\" (UID: \"7afa918d-be67-40a6-803c-d3b0ae99d815\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368209 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls\") pod \"b4750666-1362-4001-abd0-6f89964cc621\" (UID: \"b4750666-1362-4001-abd0-6f89964cc621\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368249 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368290 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368389 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle\") pod \"ce090a97-9ab6-4c40-a719-64ff2acd9778\" (UID: \"ce090a97-9ab6-4c40-a719-64ff2acd9778\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368432 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368479 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zg8nc\" (UniqueName: \"kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc\") pod \"2325ffef-9d5b-447f-b00e-3efc429acefe\" (UID: \"2325ffef-9d5b-447f-b00e-3efc429acefe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368523 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddlk9\" (UniqueName: \"kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368573 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist\") pod \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\" (UID: \"869851b9-7ffb-4af0-b166-1d8aa40a5f80\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368613 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities\") pod \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\" (UID: \"584e1f4a-8205-47d7-8efb-3afc6017c4c9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369441 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369589 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369579 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnsbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v94n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnsbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369636 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config\") pod \"736c54fe-349c-4bb9-870a-d1c1d1c03831\" (UID: \"736c54fe-349c-4bb9-870a-d1c1d1c03831\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369682 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert\") pod \"301e1965-1754-483d-b6cc-bfae7038bbca\" (UID: \"301e1965-1754-483d-b6cc-bfae7038bbca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369719 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369799 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ks6v2\" (UniqueName: \"kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2\") pod \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\" (UID: \"71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369846 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir\") pod \"c5f2bfad-70f6-4185-a3d9-81ce12720767\" (UID: \"c5f2bfad-70f6-4185-a3d9-81ce12720767\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369948 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config\") pod \"f559dfa3-3917-43a2-97f6-61ddfda10e93\" (UID: \"f559dfa3-3917-43a2-97f6-61ddfda10e93\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370062 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities\") pod \"b605f283-6f2e-42da-a838-54421690f7d0\" (UID: \"b605f283-6f2e-42da-a838-54421690f7d0\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370106 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370143 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5lgh\" (UniqueName: \"kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370177 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config\") pod \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\" (UID: \"fc8db2c7-859d-47b3-a900-2bd0c0b2973b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370213 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qgrkj\" (UniqueName: \"kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj\") pod \"42a11a02-47e1-488f-b270-2679d3298b0e\" (UID: \"42a11a02-47e1-488f-b270-2679d3298b0e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366357 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370252 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370291 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token\") pod \"593a3561-7760-45c5-8f91-5aaef7475d0f\" (UID: \"593a3561-7760-45c5-8f91-5aaef7475d0f\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370333 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert\") pod \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\" (UID: \"f65c0ac1-8bca-454d-a2e6-e35cb418beac\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370380 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config\") pod \"d19cb085-0c5b-4810-b654-ce7923221d90\" (UID: \"d19cb085-0c5b-4810-b654-ce7923221d90\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370422 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vsz9\" (UniqueName: \"kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9\") pod \"c491984c-7d4b-44aa-8c1e-d7974424fa47\" (UID: \"c491984c-7d4b-44aa-8c1e-d7974424fa47\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370463 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lcfw\" (UniqueName: \"kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw\") pod \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\" (UID: \"dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370514 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370553 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370593 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mfzkj\" (UniqueName: \"kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj\") pod \"0effdbcf-dd7d-404d-9d48-77536d665a5d\" (UID: \"0effdbcf-dd7d-404d-9d48-77536d665a5d\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370635 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5rsr\" (UniqueName: \"kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370673 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-99zj9\" (UniqueName: \"kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9\") pod \"d565531a-ff86-4608-9d19-767de01ac31b\" (UID: \"d565531a-ff86-4608-9d19-767de01ac31b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370715 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9stx\" (UniqueName: \"kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370790 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca\") pod \"9f71a554-e414-4bc3-96d2-674060397afe\" (UID: \"9f71a554-e414-4bc3-96d2-674060397afe\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370827 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370863 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access\") pod \"d7e8f42f-dc0e-424b-bb56-5ec849834888\" (UID: \"d7e8f42f-dc0e-424b-bb56-5ec849834888\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370905 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert\") pod \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\" (UID: \"7fcc6409-8a0f-44c3-89e7-5aecd7610f8a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370943 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.371031 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls\") pod \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\" (UID: \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.371111 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config\") pod \"09cfa50b-4138-4585-a53e-64dd3ab73335\" (UID: \"09cfa50b-4138-4585-a53e-64dd3ab73335\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.371155 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m26jq\" (UniqueName: \"kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.371206 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token\") pod \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\" (UID: \"20ce4d18-fe25-4696-ad7c-1bd2d6200a3e\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.371243 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.372143 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.372214 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373736 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca\") pod \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\" (UID: \"5ebfebf6-3ecd-458e-943f-bb25b52e2718\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373806 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls\") pod \"92dfbade-90b6-4169-8c07-72cff7f2c82b\" (UID: \"92dfbade-90b6-4169-8c07-72cff7f2c82b\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373836 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nspp\" (UniqueName: \"kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp\") pod \"a7a88189-c967-4640-879e-27665747f20c\" (UID: \"a7a88189-c967-4640-879e-27665747f20c\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373862 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w94wk\" (UniqueName: \"kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk\") pod \"01080b46-74f1-4191-8755-5152a57b3b25\" (UID: \"01080b46-74f1-4191-8755-5152a57b3b25\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373891 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4tqq\" (UniqueName: \"kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq\") pod \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\" (UID: \"6ee8fbd3-1f81-4666-96da-5afc70819f1a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373934 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config\") pod \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\" (UID: \"e1d2a42d-af1d-4054-9618-ab545e0ed8b7\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.366855 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config" (OuterVolumeSpecName: "config") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367457 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.367640 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp" (OuterVolumeSpecName: "tmp") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368145 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "7df94c10-441d-4386-93a6-6730fb7bcde0" (UID: "7df94c10-441d-4386-93a6-6730fb7bcde0"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368487 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368589 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz" (OuterVolumeSpecName: "kube-api-access-ws8zz") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "kube-api-access-ws8zz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368817 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "16bdd140-dce1-464c-ab47-dd5798d1d256" (UID: "16bdd140-dce1-464c-ab47-dd5798d1d256"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368761 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv" (OuterVolumeSpecName: "kube-api-access-dztfv") pod "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" (UID: "d45be74c-0d98-4d18-90e4-f7ef1b6daaf7"). InnerVolumeSpecName "kube-api-access-dztfv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368851 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6" (OuterVolumeSpecName: "kube-api-access-tkdh6") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "kube-api-access-tkdh6". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368851 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities" (OuterVolumeSpecName: "utilities") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368891 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.368985 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl" (OuterVolumeSpecName: "kube-api-access-twvbl") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "kube-api-access-twvbl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369068 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369286 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities" (OuterVolumeSpecName: "utilities") pod "584e1f4a-8205-47d7-8efb-3afc6017c4c9" (UID: "584e1f4a-8205-47d7-8efb-3afc6017c4c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369346 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd" (OuterVolumeSpecName: "kube-api-access-mjwtd") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "kube-api-access-mjwtd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369363 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs" (OuterVolumeSpecName: "kube-api-access-l87hs") pod "5ebfebf6-3ecd-458e-943f-bb25b52e2718" (UID: "5ebfebf6-3ecd-458e-943f-bb25b52e2718"). InnerVolumeSpecName "kube-api-access-l87hs". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369801 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key" (OuterVolumeSpecName: "signing-key") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369807 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem" (OuterVolumeSpecName: "ca-trust-extracted-pem") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "ca-trust-extracted-pem". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.377446 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq" (OuterVolumeSpecName: "kube-api-access-d4tqq") pod "6ee8fbd3-1f81-4666-96da-5afc70819f1a" (UID: "6ee8fbd3-1f81-4666-96da-5afc70819f1a"). InnerVolumeSpecName "kube-api-access-d4tqq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.369990 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370230 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf" (OuterVolumeSpecName: "kube-api-access-ptkcf") pod "7599e0b6-bddf-4def-b7f2-0b32206e8651" (UID: "7599e0b6-bddf-4def-b7f2-0b32206e8651"). InnerVolumeSpecName "kube-api-access-ptkcf". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370774 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl" (OuterVolumeSpecName: "kube-api-access-26xrl") pod "a208c9c2-333b-4b4a-be0d-bc32ec38a821" (UID: "a208c9c2-333b-4b4a-be0d-bc32ec38a821"). InnerVolumeSpecName "kube-api-access-26xrl". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370802 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.371119 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.370822 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config" (OuterVolumeSpecName: "config") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.371792 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities" (OuterVolumeSpecName: "utilities") pod "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" (UID: "9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.372193 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca" (OuterVolumeSpecName: "client-ca") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.372774 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373304 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373402 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.377719 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca" (OuterVolumeSpecName: "service-ca") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.377977 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m" (OuterVolumeSpecName: "kube-api-access-4hb7m") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "kube-api-access-4hb7m". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373465 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "b4750666-1362-4001-abd0-6f89964cc621" (UID: "b4750666-1362-4001-abd0-6f89964cc621"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.373908 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn" (OuterVolumeSpecName: "kube-api-access-xnxbn") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "kube-api-access-xnxbn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.374092 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.874061672 +0000 UTC m=+85.553753168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.374124 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "81e39f7b-62e4-4fc9-992a-6535ce127a02" (UID: "81e39f7b-62e4-4fc9-992a-6535ce127a02"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.374129 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378096 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.374163 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.374532 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd" (OuterVolumeSpecName: "kube-api-access-8pskd") pod "a555ff2e-0be6-46d5-897d-863bb92ae2b3" (UID: "a555ff2e-0be6-46d5-897d-863bb92ae2b3"). InnerVolumeSpecName "kube-api-access-8pskd". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.374959 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "869851b9-7ffb-4af0-b166-1d8aa40a5f80" (UID: "869851b9-7ffb-4af0-b166-1d8aa40a5f80"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378132 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login\") pod \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\" (UID: \"6edfcf45-925b-4eff-b940-95b6fc0b85d4\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378273 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle\") pod \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\" (UID: \"6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378322 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib\") pod \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\" (UID: \"af33e427-6803-48c2-a76a-dd9deb7cbf9a\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378573 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378645 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config" (OuterVolumeSpecName: "config") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378658 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert\") pod \"567683bd-0efc-4f21-b076-e28559628404\" (UID: \"567683bd-0efc-4f21-b076-e28559628404\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378847 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca" (OuterVolumeSpecName: "serviceca") pod "5ebfebf6-3ecd-458e-943f-bb25b52e2718" (UID: "5ebfebf6-3ecd-458e-943f-bb25b52e2718"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378885 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "f7e2c886-118e-43bb-bef1-c78134de392b" (UID: "f7e2c886-118e-43bb-bef1-c78134de392b"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378904 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-netns\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378951 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.378996 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379073 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-systemd-units\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.375033 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l" (OuterVolumeSpecName: "kube-api-access-sbc2l") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "kube-api-access-sbc2l". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379190 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.375054 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "736c54fe-349c-4bb9-870a-d1c1d1c03831" (UID: "736c54fe-349c-4bb9-870a-d1c1d1c03831"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379198 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-bin\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.375208 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.375661 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b" (OuterVolumeSpecName: "kube-api-access-pgx6b") pod "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" (UID: "f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4"). InnerVolumeSpecName "kube-api-access-pgx6b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.375724 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj" (OuterVolumeSpecName: "kube-api-access-mfzkj") pod "0effdbcf-dd7d-404d-9d48-77536d665a5d" (UID: "0effdbcf-dd7d-404d-9d48-77536d665a5d"). InnerVolumeSpecName "kube-api-access-mfzkj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.375801 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376299 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config" (OuterVolumeSpecName: "config") pod "fc8db2c7-859d-47b3-a900-2bd0c0b2973b" (UID: "fc8db2c7-859d-47b3-a900-2bd0c0b2973b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376447 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr" (OuterVolumeSpecName: "kube-api-access-6g4lr") pod "f7e2c886-118e-43bb-bef1-c78134de392b" (UID: "f7e2c886-118e-43bb-bef1-c78134de392b"). InnerVolumeSpecName "kube-api-access-6g4lr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376519 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376457 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379326 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq" (OuterVolumeSpecName: "kube-api-access-m26jq") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "kube-api-access-m26jq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376474 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376573 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr" (OuterVolumeSpecName: "kube-api-access-z5rsr") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "kube-api-access-z5rsr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376623 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376428 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp" (OuterVolumeSpecName: "kube-api-access-8nspp") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "kube-api-access-8nspp". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376779 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "301e1965-1754-483d-b6cc-bfae7038bbca" (UID: "301e1965-1754-483d-b6cc-bfae7038bbca"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376791 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "18f80adb-c1c3-49ba-8ee4-932c851d3897" (UID: "18f80adb-c1c3-49ba-8ee4-932c851d3897"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.376911 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "a208c9c2-333b-4b4a-be0d-bc32ec38a821" (UID: "a208c9c2-333b-4b4a-be0d-bc32ec38a821"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.377258 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f65c0ac1-8bca-454d-a2e6-e35cb418beac" (UID: "f65c0ac1-8bca-454d-a2e6-e35cb418beac"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379424 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw" (OuterVolumeSpecName: "kube-api-access-5lcfw") pod "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" (UID: "dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9"). InnerVolumeSpecName "kube-api-access-5lcfw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379109 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-bin\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.377627 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9" (OuterVolumeSpecName: "kube-api-access-ddlk9") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "kube-api-access-ddlk9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.377984 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "c5f2bfad-70f6-4185-a3d9-81ce12720767" (UID: "c5f2bfad-70f6-4185-a3d9-81ce12720767"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.377292 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "593a3561-7760-45c5-8f91-5aaef7475d0f" (UID: "593a3561-7760-45c5-8f91-5aaef7475d0f"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379496 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "6ee8fbd3-1f81-4666-96da-5afc70819f1a" (UID: "6ee8fbd3-1f81-4666-96da-5afc70819f1a"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379617 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca" (OuterVolumeSpecName: "service-ca") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379637 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "d7e8f42f-dc0e-424b-bb56-5ec849834888" (UID: "d7e8f42f-dc0e-424b-bb56-5ec849834888"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379704 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-slash\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379815 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379884 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images" (OuterVolumeSpecName: "images") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379631 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-slash\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379941 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-netd\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.379987 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config" (OuterVolumeSpecName: "config") pod "09cfa50b-4138-4585-a53e-64dd3ab73335" (UID: "09cfa50b-4138-4585-a53e-64dd3ab73335"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380044 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-netns\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380047 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-script-lib\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380087 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380093 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-netd\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380108 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-systemd-units\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.380191 5109 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380245 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380277 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380312 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.380325 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs podName:ad4ed271-a1a8-408d-b043-99cf044fb573 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:07.880292003 +0000 UTC m=+85.559983479 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs") pod "network-metrics-daemon-fp2cw" (UID: "ad4ed271-a1a8-408d-b043-99cf044fb573") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380369 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-config\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380458 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-log-socket\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380495 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-46pxq\" (UniqueName: \"kubernetes.io/projected/561da642-b358-4ddc-86ee-669164d7b3c5-kube-api-access-46pxq\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380539 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-etc-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380577 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p72wg\" (UniqueName: \"kubernetes.io/projected/61fc568b-d7c2-4633-b28c-f65428d01030-kube-api-access-p72wg\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380681 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-ovn\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380718 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380812 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x5rg8\" (UniqueName: \"kubernetes.io/projected/ad4ed271-a1a8-408d-b043-99cf044fb573-kube-api-access-x5rg8\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380907 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-var-lib-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380945 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61fc568b-d7c2-4633-b28c-f65428d01030-host\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380980 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-c4fpt\" (UniqueName: \"kubernetes.io/projected/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-kube-api-access-c4fpt\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.380975 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "567683bd-0efc-4f21-b076-e28559628404" (UID: "567683bd-0efc-4f21-b076-e28559628404"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381066 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/561da642-b358-4ddc-86ee-669164d7b3c5-ovn-node-metrics-cert\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381069 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" (UID: "7fcc6409-8a0f-44c3-89e7-5aecd7610f8a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381222 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-script-lib\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381367 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381589 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovnkube-config\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381622 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-ovn\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381710 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381994 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-etc-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.381999 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-env-overrides\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382149 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-node-log\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382180 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-ovn-kubernetes\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382204 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/61fc568b-d7c2-4633-b28c-f65428d01030-serviceca\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382277 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-log-socket\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382285 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-ovn-kubernetes\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382734 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-config\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382799 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-kubelet\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382862 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-env-overrides\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.382975 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-systemd\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383032 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-node-log\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383183 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383207 5109 reconciler_common.go:299] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7df94c10-441d-4386-93a6-6730fb7bcde0-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383228 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383249 5109 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383270 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pllx6\" (UniqueName: \"kubernetes.io/projected/81e39f7b-62e4-4fc9-992a-6535ce127a02-kube-api-access-pllx6\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383291 5109 reconciler_common.go:299] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383297 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/61fc568b-d7c2-4633-b28c-f65428d01030-serviceca\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383311 5109 reconciler_common.go:299] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383331 5109 reconciler_common.go:299] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/42a11a02-47e1-488f-b270-2679d3298b0e-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383352 5109 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/92dfbade-90b6-4169-8c07-72cff7f2c82b-tmp-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383372 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/a555ff2e-0be6-46d5-897d-863bb92ae2b3-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383391 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qqbfk\" (UniqueName: \"kubernetes.io/projected/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-kube-api-access-qqbfk\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383412 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383431 5109 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/6077b63e-53a2-4f96-9d56-1ce0324e4913-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383451 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tknt7\" (UniqueName: \"kubernetes.io/projected/584e1f4a-8205-47d7-8efb-3afc6017c4c9-kube-api-access-tknt7\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383470 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7afa918d-be67-40a6-803c-d3b0ae99d815-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383488 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383508 5109 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383528 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383534 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-env-overrides\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383547 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7599e0b6-bddf-4def-b7f2-0b32206e8651-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383679 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-var-lib-openvswitch\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383711 5109 reconciler_common.go:299] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383776 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383793 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383800 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-systemd\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383809 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9z4sw\" (UniqueName: \"kubernetes.io/projected/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-kube-api-access-9z4sw\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383808 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk" (OuterVolumeSpecName: "kube-api-access-w94wk") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "kube-api-access-w94wk". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383723 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/61fc568b-d7c2-4633-b28c-f65428d01030-host\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383854 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6rmnv\" (UniqueName: \"kubernetes.io/projected/b605f283-6f2e-42da-a838-54421690f7d0-kube-api-access-6rmnv\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383872 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383884 5109 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383896 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zsb9b\" (UniqueName: \"kubernetes.io/projected/09cfa50b-4138-4585-a53e-64dd3ab73335-kube-api-access-zsb9b\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383909 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383922 5109 reconciler_common.go:299] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383951 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-nmmzf\" (UniqueName: \"kubernetes.io/projected/7df94c10-441d-4386-93a6-6730fb7bcde0-kube-api-access-nmmzf\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383972 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383985 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.383998 5109 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/6077b63e-53a2-4f96-9d56-1ce0324e4913-tmp-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384011 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7afa918d-be67-40a6-803c-d3b0ae99d815-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384025 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7jjkz\" (UniqueName: \"kubernetes.io/projected/301e1965-1754-483d-b6cc-bfae7038bbca-kube-api-access-7jjkz\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384040 5109 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f65c0ac1-8bca-454d-a2e6-e35cb418beac-tmp-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384053 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384067 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2325ffef-9d5b-447f-b00e-3efc429acefe-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384086 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384100 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384115 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384127 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4g8ts\" (UniqueName: \"kubernetes.io/projected/92dfbade-90b6-4169-8c07-72cff7f2c82b-kube-api-access-4g8ts\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384140 5109 reconciler_common.go:299] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384153 5109 reconciler_common.go:299] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b4750666-1362-4001-abd0-6f89964cc621-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384167 5109 reconciler_common.go:299] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384180 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rzt4w\" (UniqueName: \"kubernetes.io/projected/a52afe44-fb37-46ed-a1f8-bf39727a3cbe-kube-api-access-rzt4w\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384192 5109 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/92dfbade-90b6-4169-8c07-72cff7f2c82b-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384205 5109 reconciler_common.go:299] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-certs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384217 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zth6t\" (UniqueName: \"kubernetes.io/projected/6077b63e-53a2-4f96-9d56-1ce0324e4913-kube-api-access-zth6t\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384232 5109 reconciler_common.go:299] "Volume detached for volume \"whereabouts-flatfile-configmap\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-whereabouts-flatfile-configmap\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384247 5109 reconciler_common.go:299] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384260 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a555ff2e-0be6-46d5-897d-863bb92ae2b3-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384318 5109 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384337 5109 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384353 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xxfcv\" (UniqueName: \"kubernetes.io/projected/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-kube-api-access-xxfcv\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384366 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384394 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384410 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c5f2bfad-70f6-4185-a3d9-81ce12720767-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384426 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hckvg\" (UniqueName: \"kubernetes.io/projected/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-kube-api-access-hckvg\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384439 5109 reconciler_common.go:299] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384451 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pddnv\" (UniqueName: \"kubernetes.io/projected/e093be35-bb62-4843-b2e8-094545761610-kube-api-access-pddnv\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384463 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-tkdh6\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-kube-api-access-tkdh6\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384475 5109 reconciler_common.go:299] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/16bdd140-dce1-464c-ab47-dd5798d1d256-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384487 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384499 5109 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7df94c10-441d-4386-93a6-6730fb7bcde0-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384512 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dztfv\" (UniqueName: \"kubernetes.io/projected/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-kube-api-access-dztfv\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384526 5109 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384537 5109 reconciler_common.go:299] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-key\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384549 5109 reconciler_common.go:299] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384561 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c5f2bfad-70f6-4185-a3d9-81ce12720767-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384572 5109 reconciler_common.go:299] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/18f80adb-c1c3-49ba-8ee4-932c851d3897-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384585 5109 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/9f71a554-e414-4bc3-96d2-674060397afe-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384598 5109 reconciler_common.go:299] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/c491984c-7d4b-44aa-8c1e-d7974424fa47-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384613 5109 reconciler_common.go:299] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384671 5109 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384686 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/736c54fe-349c-4bb9-870a-d1c1d1c03831-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384698 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ws8zz\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-kube-api-access-ws8zz\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384710 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d7e8f42f-dc0e-424b-bb56-5ec849834888-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384748 5109 reconciler_common.go:299] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384761 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mjwtd\" (UniqueName: \"kubernetes.io/projected/869851b9-7ffb-4af0-b166-1d8aa40a5f80-kube-api-access-mjwtd\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384774 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-twvbl\" (UniqueName: \"kubernetes.io/projected/b4750666-1362-4001-abd0-6f89964cc621-kube-api-access-twvbl\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384786 5109 reconciler_common.go:299] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384799 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ptkcf\" (UniqueName: \"kubernetes.io/projected/7599e0b6-bddf-4def-b7f2-0b32206e8651-kube-api-access-ptkcf\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384811 5109 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/f7e2c886-118e-43bb-bef1-c78134de392b-tmp-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384823 5109 reconciler_common.go:299] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384835 5109 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-ca-trust-extracted-pem\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384850 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384863 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6g4lr\" (UniqueName: \"kubernetes.io/projected/f7e2c886-118e-43bb-bef1-c78134de392b-kube-api-access-6g4lr\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384875 5109 reconciler_common.go:299] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384886 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/736c54fe-349c-4bb9-870a-d1c1d1c03831-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384899 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384911 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384924 5109 reconciler_common.go:299] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/a208c9c2-333b-4b4a-be0d-bc32ec38a821-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384936 5109 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384949 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xnxbn\" (UniqueName: \"kubernetes.io/projected/ce090a97-9ab6-4c40-a719-64ff2acd9778-kube-api-access-xnxbn\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384961 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4hb7m\" (UniqueName: \"kubernetes.io/projected/94a6e063-3d1a-4d44-875d-185291448c31-kube-api-access-4hb7m\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384976 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9e9b5059-1b3e-4067-a63d-2952cbe863af-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384989 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pgx6b\" (UniqueName: \"kubernetes.io/projected/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4-kube-api-access-pgx6b\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.384986 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385011 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-kubelet\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385003 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385325 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8pskd\" (UniqueName: \"kubernetes.io/projected/a555ff2e-0be6-46d5-897d-863bb92ae2b3-kube-api-access-8pskd\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385460 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-sbc2l\" (UniqueName: \"kubernetes.io/projected/593a3561-7760-45c5-8f91-5aaef7475d0f-kube-api-access-sbc2l\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385483 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l87hs\" (UniqueName: \"kubernetes.io/projected/5ebfebf6-3ecd-458e-943f-bb25b52e2718-kube-api-access-l87hs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385503 5109 reconciler_common.go:299] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385551 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-26xrl\" (UniqueName: \"kubernetes.io/projected/a208c9c2-333b-4b4a-be0d-bc32ec38a821-kube-api-access-26xrl\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385570 5109 reconciler_common.go:299] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385594 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385614 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b4750666-1362-4001-abd0-6f89964cc621-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385634 5109 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/301e1965-1754-483d-b6cc-bfae7038bbca-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385653 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c5f2bfad-70f6-4185-a3d9-81ce12720767-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385672 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ddlk9\" (UniqueName: \"kubernetes.io/projected/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-kube-api-access-ddlk9\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385691 5109 reconciler_common.go:299] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/869851b9-7ffb-4af0-b166-1d8aa40a5f80-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385710 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/584e1f4a-8205-47d7-8efb-3afc6017c4c9-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385729 5109 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d7e8f42f-dc0e-424b-bb56-5ec849834888-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385771 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/736c54fe-349c-4bb9-870a-d1c1d1c03831-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385790 5109 reconciler_common.go:299] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385808 5109 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c5f2bfad-70f6-4185-a3d9-81ce12720767-tmp-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385829 5109 reconciler_common.go:299] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c491984c-7d4b-44aa-8c1e-d7974424fa47-images\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385848 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385867 5109 reconciler_common.go:299] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/593a3561-7760-45c5-8f91-5aaef7475d0f-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385888 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f65c0ac1-8bca-454d-a2e6-e35cb418beac-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385907 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-5lcfw\" (UniqueName: \"kubernetes.io/projected/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-kube-api-access-5lcfw\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385926 5109 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385943 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9" (OuterVolumeSpecName: "kube-api-access-9vsz9") pod "c491984c-7d4b-44aa-8c1e-d7974424fa47" (UID: "c491984c-7d4b-44aa-8c1e-d7974424fa47"). InnerVolumeSpecName "kube-api-access-9vsz9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385982 5109 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/567683bd-0efc-4f21-b076-e28559628404-tmp-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.385985 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.386880 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps" (OuterVolumeSpecName: "kube-api-access-d7cps") pod "af41de71-79cf-4590-bbe9-9e8b848862cb" (UID: "af41de71-79cf-4590-bbe9-9e8b848862cb"). InnerVolumeSpecName "kube-api-access-d7cps". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.387777 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp" (OuterVolumeSpecName: "tmp") pod "7afa918d-be67-40a6-803c-d3b0ae99d815" (UID: "7afa918d-be67-40a6-803c-d3b0ae99d815"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.389983 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-5jnd7" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.392355 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c" (OuterVolumeSpecName: "kube-api-access-8nb9c") pod "6edfcf45-925b-4eff-b940-95b6fc0b85d4" (UID: "6edfcf45-925b-4eff-b940-95b6fc0b85d4"). InnerVolumeSpecName "kube-api-access-8nb9c". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.392863 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh" (OuterVolumeSpecName: "kube-api-access-m5lgh") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "kube-api-access-m5lgh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393105 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393154 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.386006 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mfzkj\" (UniqueName: \"kubernetes.io/projected/0effdbcf-dd7d-404d-9d48-77536d665a5d-kube-api-access-mfzkj\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393374 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-z5rsr\" (UniqueName: \"kubernetes.io/projected/af33e427-6803-48c2-a76a-dd9deb7cbf9a-kube-api-access-z5rsr\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393457 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d7e8f42f-dc0e-424b-bb56-5ec849834888-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393785 5109 reconciler_common.go:299] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393871 5109 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393959 5109 reconciler_common.go:299] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6ee8fbd3-1f81-4666-96da-5afc70819f1a-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.393839 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394022 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "149b3c48-e17c-4a66-a835-d86dabf6ff13" (UID: "149b3c48-e17c-4a66-a835-d86dabf6ff13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394136 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit" (OuterVolumeSpecName: "audit") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394047 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09cfa50b-4138-4585-a53e-64dd3ab73335-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394353 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-m26jq\" (UniqueName: \"kubernetes.io/projected/567683bd-0efc-4f21-b076-e28559628404-kube-api-access-m26jq\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394430 5109 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394521 5109 reconciler_common.go:299] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394603 5109 reconciler_common.go:299] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5ebfebf6-3ecd-458e-943f-bb25b52e2718-serviceca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394689 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8nspp\" (UniqueName: \"kubernetes.io/projected/a7a88189-c967-4640-879e-27665747f20c-kube-api-access-8nspp\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394784 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d4tqq\" (UniqueName: \"kubernetes.io/projected/6ee8fbd3-1f81-4666-96da-5afc70819f1a-kube-api-access-d4tqq\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394865 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.394936 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/567683bd-0efc-4f21-b076-e28559628404-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395024 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395102 5109 reconciler_common.go:299] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395178 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395254 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-q4smf\" (UniqueName: \"kubernetes.io/projected/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-kube-api-access-q4smf\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395409 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395501 5109 reconciler_common.go:299] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a7a88189-c967-4640-879e-27665747f20c-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395588 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f65c0ac1-8bca-454d-a2e6-e35cb418beac-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395665 5109 reconciler_common.go:299] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395752 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ftwb6\" (UniqueName: \"kubernetes.io/projected/9f71a554-e414-4bc3-96d2-674060397afe-kube-api-access-ftwb6\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395827 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395902 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6dmhf\" (UniqueName: \"kubernetes.io/projected/736c54fe-349c-4bb9-870a-d1c1d1c03831-kube-api-access-6dmhf\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.395989 5109 reconciler_common.go:299] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396075 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-94l9h\" (UniqueName: \"kubernetes.io/projected/16bdd140-dce1-464c-ab47-dd5798d1d256-kube-api-access-94l9h\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396151 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afa918d-be67-40a6-803c-d3b0ae99d815-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396212 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wj4qr\" (UniqueName: \"kubernetes.io/projected/149b3c48-e17c-4a66-a835-d86dabf6ff13-kube-api-access-wj4qr\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396278 5109 reconciler_common.go:299] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/567683bd-0efc-4f21-b076-e28559628404-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396347 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7599e0b6-bddf-4def-b7f2-0b32206e8651-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396426 5109 reconciler_common.go:299] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/301e1965-1754-483d-b6cc-bfae7038bbca-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396497 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc85e424-18b2-4924-920b-bd291a8c4b01-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396566 5109 reconciler_common.go:299] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/81e39f7b-62e4-4fc9-992a-6535ce127a02-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396642 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16bdd140-dce1-464c-ab47-dd5798d1d256-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396717 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hm9x7\" (UniqueName: \"kubernetes.io/projected/f559dfa3-3917-43a2-97f6-61ddfda10e93-kube-api-access-hm9x7\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396834 5109 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9e9b5059-1b3e-4067-a63d-2952cbe863af-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396912 5109 reconciler_common.go:299] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/fc8db2c7-859d-47b3-a900-2bd0c0b2973b-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.396987 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397067 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397147 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397224 5109 reconciler_common.go:299] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397305 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f65c0ac1-8bca-454d-a2e6-e35cb418beac-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397378 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09cfa50b-4138-4585-a53e-64dd3ab73335-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397507 5109 reconciler_common.go:299] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397605 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xfp5s\" (UniqueName: \"kubernetes.io/projected/cc85e424-18b2-4924-920b-bd291a8c4b01-kube-api-access-xfp5s\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397681 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a555ff2e-0be6-46d5-897d-863bb92ae2b3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397790 5109 reconciler_common.go:299] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397871 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f559dfa3-3917-43a2-97f6-61ddfda10e93-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397948 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398058 5109 reconciler_common.go:299] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.397956 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398167 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-grwfz\" (UniqueName: \"kubernetes.io/projected/31fa8943-81cc-4750-a0b7-0fa9ab5af883-kube-api-access-grwfz\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398362 5109 reconciler_common.go:299] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/18f80adb-c1c3-49ba-8ee4-932c851d3897-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398435 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wbmqg\" (UniqueName: \"kubernetes.io/projected/18f80adb-c1c3-49ba-8ee4-932c851d3897-kube-api-access-wbmqg\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398504 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398585 5109 reconciler_common.go:299] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-console-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398513 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config" (OuterVolumeSpecName: "config") pod "d19cb085-0c5b-4810-b654-ce7923221d90" (UID: "d19cb085-0c5b-4810-b654-ce7923221d90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398828 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp" (OuterVolumeSpecName: "tmp") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.398970 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images" (OuterVolumeSpecName: "images") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.399023 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.399584 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj" (OuterVolumeSpecName: "kube-api-access-qgrkj") pod "42a11a02-47e1-488f-b270-2679d3298b0e" (UID: "42a11a02-47e1-488f-b270-2679d3298b0e"). InnerVolumeSpecName "kube-api-access-qgrkj". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.399892 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"189d054c-6301-4cf4-ba60-da0380233d01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2wzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400012 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config" (OuterVolumeSpecName: "config") pod "01080b46-74f1-4191-8755-5152a57b3b25" (UID: "01080b46-74f1-4191-8755-5152a57b3b25"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400026 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2" (OuterVolumeSpecName: "kube-api-access-ks6v2") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "kube-api-access-ks6v2". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400236 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400357 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400673 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400777 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400883 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.400979 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.402601 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/561da642-b358-4ddc-86ee-669164d7b3c5-ovn-node-metrics-cert\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.405190 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dsgwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-5jnd7_openshift-network-operator(428b39f5-eb1c-4f65-b7a4-eeb6e84860cc): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.405869 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "92dfbade-90b6-4169-8c07-72cff7f2c82b" (UID: "92dfbade-90b6-4169-8c07-72cff7f2c82b"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.405969 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4fpt\" (UniqueName: \"kubernetes.io/projected/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-kube-api-access-c4fpt\") pod \"ovnkube-control-plane-57b78d8988-dzv7n\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.406603 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-5jnd7" podUID="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.406850 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.407015 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx" (OuterVolumeSpecName: "kube-api-access-l9stx") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "kube-api-access-l9stx". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.407205 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc" (OuterVolumeSpecName: "kube-api-access-zg8nc") pod "2325ffef-9d5b-447f-b00e-3efc429acefe" (UID: "2325ffef-9d5b-447f-b00e-3efc429acefe"). InnerVolumeSpecName "kube-api-access-zg8nc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.407278 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "0dd0fbac-8c0d-4228-8faa-abbeedabf7db" (UID: "0dd0fbac-8c0d-4228-8faa-abbeedabf7db"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.407398 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "e1d2a42d-af1d-4054-9618-ab545e0ed8b7" (UID: "e1d2a42d-af1d-4054-9618-ab545e0ed8b7"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.408520 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p72wg\" (UniqueName: \"kubernetes.io/projected/61fc568b-d7c2-4633-b28c-f65428d01030-kube-api-access-p72wg\") pod \"node-ca-xqdz5\" (UID: \"61fc568b-d7c2-4633-b28c-f65428d01030\") " pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.409068 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5rg8\" (UniqueName: \"kubernetes.io/projected/ad4ed271-a1a8-408d-b043-99cf044fb573-kube-api-access-x5rg8\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.409710 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "ce090a97-9ab6-4c40-a719-64ff2acd9778" (UID: "ce090a97-9ab6-4c40-a719-64ff2acd9778"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.409884 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities" (OuterVolumeSpecName: "utilities") pod "b605f283-6f2e-42da-a838-54421690f7d0" (UID: "b605f283-6f2e-42da-a838-54421690f7d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.410923 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-46pxq\" (UniqueName: \"kubernetes.io/projected/561da642-b358-4ddc-86ee-669164d7b3c5-kube-api-access-46pxq\") pod \"ovnkube-node-wfxmd\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.411137 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp" (OuterVolumeSpecName: "tmp") pod "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" (UID: "b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.411168 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9" (OuterVolumeSpecName: "kube-api-access-99zj9") pod "d565531a-ff86-4608-9d19-767de01ac31b" (UID: "d565531a-ff86-4608-9d19-767de01ac31b"). InnerVolumeSpecName "kube-api-access-99zj9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.411915 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "af33e427-6803-48c2-a76a-dd9deb7cbf9a" (UID: "af33e427-6803-48c2-a76a-dd9deb7cbf9a"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.412093 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.412452 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" (UID: "20ce4d18-fe25-4696-ad7c-1bd2d6200a3e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.412537 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.412940 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "f559dfa3-3917-43a2-97f6-61ddfda10e93" (UID: "f559dfa3-3917-43a2-97f6-61ddfda10e93"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.413730 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" (UID: "6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.413827 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "a7a88189-c967-4640-879e-27665747f20c" (UID: "a7a88189-c967-4640-879e-27665747f20c"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.414218 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9f71a554-e414-4bc3-96d2-674060397afe" (UID: "9f71a554-e414-4bc3-96d2-674060397afe"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.423565 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad4ed271-a1a8-408d-b043-99cf044fb573\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fp2cw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.424711 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" (UID: "71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.426973 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "31fa8943-81cc-4750-a0b7-0fa9ab5af883" (UID: "31fa8943-81cc-4750-a0b7-0fa9ab5af883"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.432689 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.438153 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.442047 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94a6e063-3d1a-4d44-875d-185291448c31" (UID: "94a6e063-3d1a-4d44-875d-185291448c31"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.443303 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.454176 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnsbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v94n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnsbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.466886 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-rnsbf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.470680 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"189d054c-6301-4cf4-ba60-da0380233d01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2wzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.478217 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-hx2kb" Dec 11 16:54:07 crc kubenswrapper[5109]: W1211 16:54:07.481223 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod305d0e9b_83ad_424c_a5a5_c4ba12d3ae45.slice/crio-afac7c21d9f5b75a22cbfc903eaebec8655f4fb4c113b38dfa91bd6c347cc01b WatchSource:0}: Error finding container afac7c21d9f5b75a22cbfc903eaebec8655f4fb4c113b38dfa91bd6c347cc01b: Status 404 returned error can't find the container with id afac7c21d9f5b75a22cbfc903eaebec8655f4fb4c113b38dfa91bd6c347cc01b Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.481716 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.484809 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e,Command:[/bin/bash -c #!/bin/bash Dec 11 16:54:07 crc kubenswrapper[5109]: set -uo pipefail Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Dec 11 16:54:07 crc kubenswrapper[5109]: HOSTS_FILE="/etc/hosts" Dec 11 16:54:07 crc kubenswrapper[5109]: TEMP_FILE="/tmp/hosts.tmp" Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: IFS=', ' read -r -a services <<< "${SERVICES}" Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: # Make a temporary file with the old hosts file's attributes. Dec 11 16:54:07 crc kubenswrapper[5109]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Dec 11 16:54:07 crc kubenswrapper[5109]: echo "Failed to preserve hosts file. Exiting." Dec 11 16:54:07 crc kubenswrapper[5109]: exit 1 Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: while true; do Dec 11 16:54:07 crc kubenswrapper[5109]: declare -A svc_ips Dec 11 16:54:07 crc kubenswrapper[5109]: for svc in "${services[@]}"; do Dec 11 16:54:07 crc kubenswrapper[5109]: # Fetch service IP from cluster dns if present. We make several tries Dec 11 16:54:07 crc kubenswrapper[5109]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Dec 11 16:54:07 crc kubenswrapper[5109]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Dec 11 16:54:07 crc kubenswrapper[5109]: # support UDP loadbalancers and require reaching DNS through TCP. Dec 11 16:54:07 crc kubenswrapper[5109]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Dec 11 16:54:07 crc kubenswrapper[5109]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Dec 11 16:54:07 crc kubenswrapper[5109]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Dec 11 16:54:07 crc kubenswrapper[5109]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Dec 11 16:54:07 crc kubenswrapper[5109]: for i in ${!cmds[*]} Dec 11 16:54:07 crc kubenswrapper[5109]: do Dec 11 16:54:07 crc kubenswrapper[5109]: ips=($(eval "${cmds[i]}")) Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: svc_ips["${svc}"]="${ips[@]}" Dec 11 16:54:07 crc kubenswrapper[5109]: break Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: # Update /etc/hosts only if we get valid service IPs Dec 11 16:54:07 crc kubenswrapper[5109]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Dec 11 16:54:07 crc kubenswrapper[5109]: # Stale entries could exist in /etc/hosts if the service is deleted Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -n "${svc_ips[*]-}" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Dec 11 16:54:07 crc kubenswrapper[5109]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Dec 11 16:54:07 crc kubenswrapper[5109]: # Only continue rebuilding the hosts entries if its original content is preserved Dec 11 16:54:07 crc kubenswrapper[5109]: sleep 60 & wait Dec 11 16:54:07 crc kubenswrapper[5109]: continue Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: # Append resolver entries for services Dec 11 16:54:07 crc kubenswrapper[5109]: rc=0 Dec 11 16:54:07 crc kubenswrapper[5109]: for svc in "${!svc_ips[@]}"; do Dec 11 16:54:07 crc kubenswrapper[5109]: for ip in ${svc_ips[${svc}]}; do Dec 11 16:54:07 crc kubenswrapper[5109]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ $rc -ne 0 ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: sleep 60 & wait Dec 11 16:54:07 crc kubenswrapper[5109]: continue Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Dec 11 16:54:07 crc kubenswrapper[5109]: # Replace /etc/hosts with our modified version if needed Dec 11 16:54:07 crc kubenswrapper[5109]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Dec 11 16:54:07 crc kubenswrapper[5109]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: sleep 60 & wait Dec 11 16:54:07 crc kubenswrapper[5109]: unset svc_ips Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6v94n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-rnsbf_openshift-dns(305d0e9b-83ad-424c-a5a5-c4ba12d3ae45): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.485989 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-rnsbf" podUID="305d0e9b-83ad-424c-a5a5-c4ba12d3ae45" Dec 11 16:54:07 crc kubenswrapper[5109]: W1211 16:54:07.489766 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c45b4ae_02aa_4614_977b_544a9e303bdf.slice/crio-419ecb6687f142b181cf0a23840a7b085438301c10b68baf51ac446dd3b597e1 WatchSource:0}: Error finding container 419ecb6687f142b181cf0a23840a7b085438301c10b68baf51ac446dd3b597e1: Status 404 returned error can't find the container with id 419ecb6687f142b181cf0a23840a7b085438301c10b68baf51ac446dd3b597e1 Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.491546 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad4ed271-a1a8-408d-b043-99cf044fb573\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fp2cw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.492308 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Dec 11 16:54:07 crc kubenswrapper[5109]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Dec 11 16:54:07 crc kubenswrapper[5109]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nbrkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-hx2kb_openshift-multus(7c45b4ae-02aa-4614-977b-544a9e303bdf): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.492731 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.493668 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-hx2kb" podUID="7c45b4ae-02aa-4614-977b-544a9e303bdf" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499134 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499170 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499221 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499260 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499293 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/31fa8943-81cc-4750-a0b7-0fa9ab5af883-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499304 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499315 5109 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9e9b5059-1b3e-4067-a63d-2952cbe863af-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499326 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/149b3c48-e17c-4a66-a835-d86dabf6ff13-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499334 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01080b46-74f1-4191-8755-5152a57b3b25-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499343 5109 reconciler_common.go:299] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/0dd0fbac-8c0d-4228-8faa-abbeedabf7db-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499352 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d7cps\" (UniqueName: \"kubernetes.io/projected/af41de71-79cf-4590-bbe9-9e8b848862cb-kube-api-access-d7cps\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499361 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499370 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-8nb9c\" (UniqueName: \"kubernetes.io/projected/6edfcf45-925b-4eff-b940-95b6fc0b85d4-kube-api-access-8nb9c\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499379 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d565531a-ff86-4608-9d19-767de01ac31b-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499387 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499396 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94a6e063-3d1a-4d44-875d-185291448c31-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499406 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19cb085-0c5b-4810-b654-ce7923221d90-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499415 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6edfcf45-925b-4eff-b940-95b6fc0b85d4-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499425 5109 reconciler_common.go:299] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499433 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2325ffef-9d5b-447f-b00e-3efc429acefe-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499441 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499450 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/7afa918d-be67-40a6-803c-d3b0ae99d815-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499469 5109 reconciler_common.go:299] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/d565531a-ff86-4608-9d19-767de01ac31b-images\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499478 5109 reconciler_common.go:299] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/ce090a97-9ab6-4c40-a719-64ff2acd9778-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499486 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zg8nc\" (UniqueName: \"kubernetes.io/projected/2325ffef-9d5b-447f-b00e-3efc429acefe-kube-api-access-zg8nc\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499495 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499503 5109 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9e9b5059-1b3e-4067-a63d-2952cbe863af-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499511 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-ks6v2\" (UniqueName: \"kubernetes.io/projected/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a-kube-api-access-ks6v2\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499520 5109 reconciler_common.go:299] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f559dfa3-3917-43a2-97f6-61ddfda10e93-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499529 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b605f283-6f2e-42da-a838-54421690f7d0-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499537 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-m5lgh\" (UniqueName: \"kubernetes.io/projected/d19cb085-0c5b-4810-b654-ce7923221d90-kube-api-access-m5lgh\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499546 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-qgrkj\" (UniqueName: \"kubernetes.io/projected/42a11a02-47e1-488f-b270-2679d3298b0e-kube-api-access-qgrkj\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499554 5109 reconciler_common.go:299] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-audit\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499561 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d19cb085-0c5b-4810-b654-ce7923221d90-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499570 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9vsz9\" (UniqueName: \"kubernetes.io/projected/c491984c-7d4b-44aa-8c1e-d7974424fa47-kube-api-access-9vsz9\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499578 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-99zj9\" (UniqueName: \"kubernetes.io/projected/d565531a-ff86-4608-9d19-767de01ac31b-kube-api-access-99zj9\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499586 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l9stx\" (UniqueName: \"kubernetes.io/projected/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-kube-api-access-l9stx\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499594 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9f71a554-e414-4bc3-96d2-674060397afe-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499602 5109 reconciler_common.go:299] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a7a88189-c967-4640-879e-27665747f20c-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499610 5109 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499618 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01080b46-74f1-4191-8755-5152a57b3b25-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499626 5109 reconciler_common.go:299] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/92dfbade-90b6-4169-8c07-72cff7f2c82b-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499635 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-w94wk\" (UniqueName: \"kubernetes.io/projected/01080b46-74f1-4191-8755-5152a57b3b25-kube-api-access-w94wk\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499645 5109 reconciler_common.go:299] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e1d2a42d-af1d-4054-9618-ab545e0ed8b7-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499653 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.499661 5109 reconciler_common.go:299] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/af33e427-6803-48c2-a76a-dd9deb7cbf9a-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.499772 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.499785 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.499795 5109 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.499838 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:08.499824708 +0000 UTC m=+86.179516174 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500016 5109 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500104 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500116 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500123 5109 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500136 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:08.500106785 +0000 UTC m=+86.179798291 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500165 5109 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500228 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:08.500161006 +0000 UTC m=+86.179852512 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.500269 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:08.500254028 +0000 UTC m=+86.179945534 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.502127 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25a760d4-55be-452a-b344-fe5e4ae5b5a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c10e7a3d56908a1837ca1b29a944fbe9febec4030a733cd53c0ecda0177091c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b0e470f587ad2b1f84cb1100ce6155dbe9ba3e4a36c01970e843d5cdf38b3d56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c37b01eb1234360d043eff79f1eb374f5d95c1132e416be3c05ad3fd9fdfc36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.502924 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.502976 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.502997 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.503022 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.503047 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.507669 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l6lpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-c2wzl_openshift-multus(189d054c-6301-4cf4-ba60-da0380233d01): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.508832 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" podUID="189d054c-6301-4cf4-ba60-da0380233d01" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.510463 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6525f27e-56ac-4b08-ac25-b854b4edf7fc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6ab4918fcfdbfaadb2441c5d160818ac2fd2f2d718e73fc0ef7d9d62dc004b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.518623 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1f09ed1-77b5-404d-abd1-44afa5fab784\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xgkvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.518720 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.527570 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.527572 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xqdz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fc568b-d7c2-4633-b28c-f65428d01030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p72wg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xqdz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: W1211 16:54:07.540497 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod561da642_b358_4ddc_86ee_669164d7b3c5.slice/crio-c36f1d88fedba6cd1a52d696a396a0a6177473d575d2e49ea1426299b95ac0f2 WatchSource:0}: Error finding container c36f1d88fedba6cd1a52d696a396a0a6177473d575d2e49ea1426299b95ac0f2: Status 404 returned error can't find the container with id c36f1d88fedba6cd1a52d696a396a0a6177473d575d2e49ea1426299b95ac0f2 Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.541812 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13d2072-a409-4658-aafa-a35c7c947e4e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://b9608c14d8ea51cb0920addf885ee63ff3e4c8a81ab46a226edebd9a0bb6a399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://58b2adf4e750dc18bc35c0aea18178cf6daed481b29ea518459defa47769e5d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://1cbeee4ade3904eb92ab061e705d6387d66e4f501d3ee96e67260accf10d2acb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.544643 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Dec 11 16:54:07 crc kubenswrapper[5109]: apiVersion: v1 Dec 11 16:54:07 crc kubenswrapper[5109]: clusters: Dec 11 16:54:07 crc kubenswrapper[5109]: - cluster: Dec 11 16:54:07 crc kubenswrapper[5109]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Dec 11 16:54:07 crc kubenswrapper[5109]: server: https://api-int.crc.testing:6443 Dec 11 16:54:07 crc kubenswrapper[5109]: name: default-cluster Dec 11 16:54:07 crc kubenswrapper[5109]: contexts: Dec 11 16:54:07 crc kubenswrapper[5109]: - context: Dec 11 16:54:07 crc kubenswrapper[5109]: cluster: default-cluster Dec 11 16:54:07 crc kubenswrapper[5109]: namespace: default Dec 11 16:54:07 crc kubenswrapper[5109]: user: default-auth Dec 11 16:54:07 crc kubenswrapper[5109]: name: default-context Dec 11 16:54:07 crc kubenswrapper[5109]: current-context: default-context Dec 11 16:54:07 crc kubenswrapper[5109]: kind: Config Dec 11 16:54:07 crc kubenswrapper[5109]: preferences: {} Dec 11 16:54:07 crc kubenswrapper[5109]: users: Dec 11 16:54:07 crc kubenswrapper[5109]: - name: default-auth Dec 11 16:54:07 crc kubenswrapper[5109]: user: Dec 11 16:54:07 crc kubenswrapper[5109]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Dec 11 16:54:07 crc kubenswrapper[5109]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Dec 11 16:54:07 crc kubenswrapper[5109]: EOF Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-46pxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-wfxmd_openshift-ovn-kubernetes(561da642-b358-4ddc-86ee-669164d7b3c5): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.547171 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5,Command:[/bin/bash -c #!/bin/bash Dec 11 16:54:07 crc kubenswrapper[5109]: set -euo pipefail Dec 11 16:54:07 crc kubenswrapper[5109]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Dec 11 16:54:07 crc kubenswrapper[5109]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Dec 11 16:54:07 crc kubenswrapper[5109]: # As the secret mount is optional we must wait for the files to be present. Dec 11 16:54:07 crc kubenswrapper[5109]: # The service is created in monitor.yaml and this is created in sdn.yaml. Dec 11 16:54:07 crc kubenswrapper[5109]: TS=$(date +%s) Dec 11 16:54:07 crc kubenswrapper[5109]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Dec 11 16:54:07 crc kubenswrapper[5109]: HAS_LOGGED_INFO=0 Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: log_missing_certs(){ Dec 11 16:54:07 crc kubenswrapper[5109]: CUR_TS=$(date +%s) Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Dec 11 16:54:07 crc kubenswrapper[5109]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Dec 11 16:54:07 crc kubenswrapper[5109]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Dec 11 16:54:07 crc kubenswrapper[5109]: HAS_LOGGED_INFO=1 Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: } Dec 11 16:54:07 crc kubenswrapper[5109]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Dec 11 16:54:07 crc kubenswrapper[5109]: log_missing_certs Dec 11 16:54:07 crc kubenswrapper[5109]: sleep 5 Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/kube-rbac-proxy \ Dec 11 16:54:07 crc kubenswrapper[5109]: --logtostderr \ Dec 11 16:54:07 crc kubenswrapper[5109]: --secure-listen-address=:9108 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Dec 11 16:54:07 crc kubenswrapper[5109]: --upstream=http://127.0.0.1:29108/ \ Dec 11 16:54:07 crc kubenswrapper[5109]: --tls-private-key-file=${TLS_PK} \ Dec 11 16:54:07 crc kubenswrapper[5109]: --tls-cert-file=${TLS_CERT} Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c4fpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-57b78d8988-dzv7n_openshift-ovn-kubernetes(5f6bde63-48d8-4726-b7d6-0f36f7cf47f9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.547497 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.552335 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c set -xe Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ -f "/env/_master" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: source "/env/_master" Dec 11 16:54:07 crc kubenswrapper[5109]: set +o allexport Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v4_join_subnet_opt= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v6_join_subnet_opt= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v4_transit_switch_subnet_opt= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v6_transit_switch_subnet_opt= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: dns_name_resolver_enabled_flag= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: persistent_ips_enabled_flag="--enable-persistent-ips" Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: # This is needed so that converting clusters from GA to TP Dec 11 16:54:07 crc kubenswrapper[5109]: # will rollout control plane pods as well Dec 11 16:54:07 crc kubenswrapper[5109]: network_segmentation_enabled_flag= Dec 11 16:54:07 crc kubenswrapper[5109]: multi_network_enabled_flag= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "true" == "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: multi_network_enabled_flag="--enable-multi-network" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "true" == "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "true" != "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: multi_network_enabled_flag="--enable-multi-network" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: network_segmentation_enabled_flag="--enable-network-segmentation" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: route_advertisements_enable_flag= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: route_advertisements_enable_flag="--enable-route-advertisements" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: preconfigured_udn_addresses_enable_flag= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: preconfigured_udn_addresses_enable_flag="--enable-preconfigured-udn-addresses" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: # Enable multi-network policy if configured (control-plane always full mode) Dec 11 16:54:07 crc kubenswrapper[5109]: multi_network_policy_enabled_flag= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: multi_network_policy_enabled_flag="--enable-multi-networkpolicy" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: # Enable admin network policy if configured (control-plane always full mode) Dec 11 16:54:07 crc kubenswrapper[5109]: admin_network_policy_enabled_flag= Dec 11 16:54:07 crc kubenswrapper[5109]: if [[ "true" == "true" ]]; then Dec 11 16:54:07 crc kubenswrapper[5109]: admin_network_policy_enabled_flag="--enable-admin-network-policy" Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: if [ "shared" == "shared" ]; then Dec 11 16:54:07 crc kubenswrapper[5109]: gateway_mode_flags="--gateway-mode shared" Dec 11 16:54:07 crc kubenswrapper[5109]: elif [ "shared" == "local" ]; then Dec 11 16:54:07 crc kubenswrapper[5109]: gateway_mode_flags="--gateway-mode local" Dec 11 16:54:07 crc kubenswrapper[5109]: else Dec 11 16:54:07 crc kubenswrapper[5109]: echo "Invalid OVN_GATEWAY_MODE: \"shared\". Must be \"local\" or \"shared\"." Dec 11 16:54:07 crc kubenswrapper[5109]: exit 1 Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: Dec 11 16:54:07 crc kubenswrapper[5109]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Dec 11 16:54:07 crc kubenswrapper[5109]: exec /usr/bin/ovnkube \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-interconnect \ Dec 11 16:54:07 crc kubenswrapper[5109]: --init-cluster-manager "${K8S_NODE}" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --config-file=/run/ovnkube-config/ovnkube.conf \ Dec 11 16:54:07 crc kubenswrapper[5109]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --metrics-bind-address "127.0.0.1:29108" \ Dec 11 16:54:07 crc kubenswrapper[5109]: --metrics-enable-pprof \ Dec 11 16:54:07 crc kubenswrapper[5109]: --metrics-enable-config-duration \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${ovn_v4_join_subnet_opt} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${ovn_v6_join_subnet_opt} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${ovn_v4_transit_switch_subnet_opt} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${ovn_v6_transit_switch_subnet_opt} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${dns_name_resolver_enabled_flag} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${persistent_ips_enabled_flag} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${multi_network_enabled_flag} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${network_segmentation_enabled_flag} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${gateway_mode_flags} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${route_advertisements_enable_flag} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${preconfigured_udn_addresses_enable_flag} \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-egress-ip=true \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-egress-firewall=true \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-egress-qos=true \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-egress-service=true \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-multicast \ Dec 11 16:54:07 crc kubenswrapper[5109]: --enable-multi-external-gateway=true \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${multi_network_policy_enabled_flag} \ Dec 11 16:54:07 crc kubenswrapper[5109]: ${admin_network_policy_enabled_flag} Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c4fpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-57b78d8988-dzv7n_openshift-ovn-kubernetes(5f6bde63-48d8-4726-b7d6-0f36f7cf47f9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.553985 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.556233 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.565710 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.575493 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-hx2kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c45b4ae-02aa-4614-977b-544a9e303bdf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbrkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hx2kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.576398 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xqdz5" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.590020 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:07 crc kubenswrapper[5109]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Dec 11 16:54:07 crc kubenswrapper[5109]: while [ true ]; Dec 11 16:54:07 crc kubenswrapper[5109]: do Dec 11 16:54:07 crc kubenswrapper[5109]: for f in $(ls /tmp/serviceca); do Dec 11 16:54:07 crc kubenswrapper[5109]: echo $f Dec 11 16:54:07 crc kubenswrapper[5109]: ca_file_path="/tmp/serviceca/${f}" Dec 11 16:54:07 crc kubenswrapper[5109]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Dec 11 16:54:07 crc kubenswrapper[5109]: reg_dir_path="/etc/docker/certs.d/${f}" Dec 11 16:54:07 crc kubenswrapper[5109]: if [ -e "${reg_dir_path}" ]; then Dec 11 16:54:07 crc kubenswrapper[5109]: cp -u $ca_file_path $reg_dir_path/ca.crt Dec 11 16:54:07 crc kubenswrapper[5109]: else Dec 11 16:54:07 crc kubenswrapper[5109]: mkdir $reg_dir_path Dec 11 16:54:07 crc kubenswrapper[5109]: cp $ca_file_path $reg_dir_path/ca.crt Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: for d in $(ls /etc/docker/certs.d); do Dec 11 16:54:07 crc kubenswrapper[5109]: echo $d Dec 11 16:54:07 crc kubenswrapper[5109]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Dec 11 16:54:07 crc kubenswrapper[5109]: reg_conf_path="/tmp/serviceca/${dp}" Dec 11 16:54:07 crc kubenswrapper[5109]: if [ ! -e "${reg_conf_path}" ]; then Dec 11 16:54:07 crc kubenswrapper[5109]: rm -rf /etc/docker/certs.d/$d Dec 11 16:54:07 crc kubenswrapper[5109]: fi Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: sleep 60 & wait ${!} Dec 11 16:54:07 crc kubenswrapper[5109]: done Dec 11 16:54:07 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p72wg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-xqdz5_openshift-image-registry(61fc568b-d7c2-4633-b28c-f65428d01030): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:07 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.592306 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-xqdz5" podUID="61fc568b-d7c2-4633-b28c-f65428d01030" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.600172 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5vmt7\" (UniqueName: \"kubernetes.io/projected/c1f09ed1-77b5-404d-abd1-44afa5fab784-kube-api-access-5vmt7\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.600216 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1f09ed1-77b5-404d-abd1-44afa5fab784-proxy-tls\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.600236 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1f09ed1-77b5-404d-abd1-44afa5fab784-mcd-auth-proxy-config\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.600826 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c1f09ed1-77b5-404d-abd1-44afa5fab784-mcd-auth-proxy-config\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.602565 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561da642-b358-4ddc-86ee-669164d7b3c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wfxmd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.603379 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c1f09ed1-77b5-404d-abd1-44afa5fab784-proxy-tls\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.603636 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vmt7\" (UniqueName: \"kubernetes.io/projected/c1f09ed1-77b5-404d-abd1-44afa5fab784-kube-api-access-5vmt7\") pod \"machine-config-daemon-xgkvv\" (UID: \"c1f09ed1-77b5-404d-abd1-44afa5fab784\") " pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.605592 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.605650 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.605669 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.605692 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.605713 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.613010 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-dzv7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.694803 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d83f88-292e-48c6-a9d2-1613e83851e6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://0ec3b28b6104ac258b2e97c9f10ec3d55ea93bbf9d9165fdce21648201c85e9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://352c8d599c93e387e36a5488dcc02f39161e4d8c0a9fbf050e4e19c3c1a4fd92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://445fc06154e726c5f8c28bf3f36d0bc02dde5ccc34cfefcb928cbee1eb9c0d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ca35d4089afda84920619ba9d77e28fdb8bf432b26bf3cb949151fce931db7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://a3b479ab0d924d9fbb549995801bad80a45ec723aff9b99d07b5d67b37675cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.708764 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.708839 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.708858 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.708884 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.708903 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.716292 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c5b2f1d-3171-42d0-8861-bdb91323d73b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T16:53:38Z\\\",\\\"message\\\":\\\"vvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nW1211 16:53:37.766372 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1211 16:53:37.766542 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI1211 16:53:37.769894 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1407815461/tls.crt::/tmp/serving-cert-1407815461/tls.key\\\\\\\"\\\\nI1211 16:53:38.166632 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 16:53:38.168550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 16:53:38.168573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 16:53:38.168605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 16:53:38.168612 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 16:53:38.173779 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 16:53:38.173817 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173826 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 16:53:38.173842 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 16:53:38.173848 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 16:53:38.173853 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 16:53:38.173934 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 16:53:38.175696 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T16:53:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.734942 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.761014 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:54:07 crc kubenswrapper[5109]: W1211 16:54:07.775726 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1f09ed1_77b5_404d_abd1_44afa5fab784.slice/crio-eac975436973d975d59dc42980083caa4c3b7d61681f938761e71c447d399da3 WatchSource:0}: Error finding container eac975436973d975d59dc42980083caa4c3b7d61681f938761e71c447d399da3: Status 404 returned error can't find the container with id eac975436973d975d59dc42980083caa4c3b7d61681f938761e71c447d399da3 Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.779527 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.20.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vmt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-xgkvv_openshift-machine-config-operator(c1f09ed1-77b5-404d-abd1-44afa5fab784): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.782389 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vmt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-xgkvv_openshift-machine-config-operator(c1f09ed1-77b5-404d-abd1-44afa5fab784): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.784029 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.811640 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.811700 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.811724 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.811780 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.811797 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.905040 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.905464 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.905489 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:08.90545282 +0000 UTC m=+86.585144326 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.905600 5109 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: E1211 16:54:07.905698 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs podName:ad4ed271-a1a8-408d-b043-99cf044fb573 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:08.905672856 +0000 UTC m=+86.585364362 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs") pod "network-metrics-daemon-fp2cw" (UID: "ad4ed271-a1a8-408d-b043-99cf044fb573") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.914008 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.914172 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.914202 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.914228 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:07 crc kubenswrapper[5109]: I1211 16:54:07.914248 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:07Z","lastTransitionTime":"2025-12-11T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.016477 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.016516 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.016525 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.016539 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.016549 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.119415 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.119481 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.119500 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.119525 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.119544 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.222526 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.222588 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.222608 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.222634 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.222653 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.271689 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"c36f1d88fedba6cd1a52d696a396a0a6177473d575d2e49ea1426299b95ac0f2"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.272981 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"eac975436973d975d59dc42980083caa4c3b7d61681f938761e71c447d399da3"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.275011 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" event={"ID":"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9","Type":"ContainerStarted","Data":"f40c60bdf53415aae65189c284cbebca1497a09cbfeb3296bd29d5af60ad6ecb"} Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.275454 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.20.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vmt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-xgkvv_openshift-machine-config-operator(c1f09ed1-77b5-404d-abd1-44afa5fab784): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.276630 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" event={"ID":"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc","Type":"ContainerStarted","Data":"4c047e9c12956e1bf77045f49d145e688658ab492332130f8c49c87c56419551"} Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.276651 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:08 crc kubenswrapper[5109]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Dec 11 16:54:08 crc kubenswrapper[5109]: apiVersion: v1 Dec 11 16:54:08 crc kubenswrapper[5109]: clusters: Dec 11 16:54:08 crc kubenswrapper[5109]: - cluster: Dec 11 16:54:08 crc kubenswrapper[5109]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Dec 11 16:54:08 crc kubenswrapper[5109]: server: https://api-int.crc.testing:6443 Dec 11 16:54:08 crc kubenswrapper[5109]: name: default-cluster Dec 11 16:54:08 crc kubenswrapper[5109]: contexts: Dec 11 16:54:08 crc kubenswrapper[5109]: - context: Dec 11 16:54:08 crc kubenswrapper[5109]: cluster: default-cluster Dec 11 16:54:08 crc kubenswrapper[5109]: namespace: default Dec 11 16:54:08 crc kubenswrapper[5109]: user: default-auth Dec 11 16:54:08 crc kubenswrapper[5109]: name: default-context Dec 11 16:54:08 crc kubenswrapper[5109]: current-context: default-context Dec 11 16:54:08 crc kubenswrapper[5109]: kind: Config Dec 11 16:54:08 crc kubenswrapper[5109]: preferences: {} Dec 11 16:54:08 crc kubenswrapper[5109]: users: Dec 11 16:54:08 crc kubenswrapper[5109]: - name: default-auth Dec 11 16:54:08 crc kubenswrapper[5109]: user: Dec 11 16:54:08 crc kubenswrapper[5109]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Dec 11 16:54:08 crc kubenswrapper[5109]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Dec 11 16:54:08 crc kubenswrapper[5109]: EOF Dec 11 16:54:08 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-46pxq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-wfxmd_openshift-ovn-kubernetes(561da642-b358-4ddc-86ee-669164d7b3c5): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:08 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.277914 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5vmt7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-xgkvv_openshift-machine-config-operator(c1f09ed1-77b5-404d-abd1-44afa5fab784): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.278174 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.279531 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:08 crc kubenswrapper[5109]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5,Command:[/bin/bash -c #!/bin/bash Dec 11 16:54:08 crc kubenswrapper[5109]: set -euo pipefail Dec 11 16:54:08 crc kubenswrapper[5109]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Dec 11 16:54:08 crc kubenswrapper[5109]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Dec 11 16:54:08 crc kubenswrapper[5109]: # As the secret mount is optional we must wait for the files to be present. Dec 11 16:54:08 crc kubenswrapper[5109]: # The service is created in monitor.yaml and this is created in sdn.yaml. Dec 11 16:54:08 crc kubenswrapper[5109]: TS=$(date +%s) Dec 11 16:54:08 crc kubenswrapper[5109]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Dec 11 16:54:08 crc kubenswrapper[5109]: HAS_LOGGED_INFO=0 Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: log_missing_certs(){ Dec 11 16:54:08 crc kubenswrapper[5109]: CUR_TS=$(date +%s) Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Dec 11 16:54:08 crc kubenswrapper[5109]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Dec 11 16:54:08 crc kubenswrapper[5109]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Dec 11 16:54:08 crc kubenswrapper[5109]: HAS_LOGGED_INFO=1 Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: } Dec 11 16:54:08 crc kubenswrapper[5109]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Dec 11 16:54:08 crc kubenswrapper[5109]: log_missing_certs Dec 11 16:54:08 crc kubenswrapper[5109]: sleep 5 Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Dec 11 16:54:08 crc kubenswrapper[5109]: exec /usr/bin/kube-rbac-proxy \ Dec 11 16:54:08 crc kubenswrapper[5109]: --logtostderr \ Dec 11 16:54:08 crc kubenswrapper[5109]: --secure-listen-address=:9108 \ Dec 11 16:54:08 crc kubenswrapper[5109]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Dec 11 16:54:08 crc kubenswrapper[5109]: --upstream=http://127.0.0.1:29108/ \ Dec 11 16:54:08 crc kubenswrapper[5109]: --tls-private-key-file=${TLS_PK} \ Dec 11 16:54:08 crc kubenswrapper[5109]: --tls-cert-file=${TLS_CERT} Dec 11 16:54:08 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c4fpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-57b78d8988-dzv7n_openshift-ovn-kubernetes(5f6bde63-48d8-4726-b7d6-0f36f7cf47f9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:08 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.279646 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.282205 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerStarted","Data":"043fedb4871ecf583cc428fb65aa96153f48f21e578feb9feb252292f3c08a57"} Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.283018 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dsgwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-5jnd7_openshift-network-operator(428b39f5-eb1c-4f65-b7a4-eeb6e84860cc): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.283375 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:08 crc kubenswrapper[5109]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122,Command:[/bin/bash -c set -xe Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ -f "/env/_master" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: set -o allexport Dec 11 16:54:08 crc kubenswrapper[5109]: source "/env/_master" Dec 11 16:54:08 crc kubenswrapper[5109]: set +o allexport Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v4_join_subnet_opt= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v6_join_subnet_opt= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v4_transit_switch_subnet_opt= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v6_transit_switch_subnet_opt= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "" != "" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: dns_name_resolver_enabled_flag= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: persistent_ips_enabled_flag="--enable-persistent-ips" Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: # This is needed so that converting clusters from GA to TP Dec 11 16:54:08 crc kubenswrapper[5109]: # will rollout control plane pods as well Dec 11 16:54:08 crc kubenswrapper[5109]: network_segmentation_enabled_flag= Dec 11 16:54:08 crc kubenswrapper[5109]: multi_network_enabled_flag= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "true" == "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: multi_network_enabled_flag="--enable-multi-network" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "true" == "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "true" != "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: multi_network_enabled_flag="--enable-multi-network" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: network_segmentation_enabled_flag="--enable-network-segmentation" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: route_advertisements_enable_flag= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: route_advertisements_enable_flag="--enable-route-advertisements" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: preconfigured_udn_addresses_enable_flag= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: preconfigured_udn_addresses_enable_flag="--enable-preconfigured-udn-addresses" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: # Enable multi-network policy if configured (control-plane always full mode) Dec 11 16:54:08 crc kubenswrapper[5109]: multi_network_policy_enabled_flag= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "false" == "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: multi_network_policy_enabled_flag="--enable-multi-networkpolicy" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: # Enable admin network policy if configured (control-plane always full mode) Dec 11 16:54:08 crc kubenswrapper[5109]: admin_network_policy_enabled_flag= Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "true" == "true" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: admin_network_policy_enabled_flag="--enable-admin-network-policy" Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: if [ "shared" == "shared" ]; then Dec 11 16:54:08 crc kubenswrapper[5109]: gateway_mode_flags="--gateway-mode shared" Dec 11 16:54:08 crc kubenswrapper[5109]: elif [ "shared" == "local" ]; then Dec 11 16:54:08 crc kubenswrapper[5109]: gateway_mode_flags="--gateway-mode local" Dec 11 16:54:08 crc kubenswrapper[5109]: else Dec 11 16:54:08 crc kubenswrapper[5109]: echo "Invalid OVN_GATEWAY_MODE: \"shared\". Must be \"local\" or \"shared\"." Dec 11 16:54:08 crc kubenswrapper[5109]: exit 1 Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Dec 11 16:54:08 crc kubenswrapper[5109]: exec /usr/bin/ovnkube \ Dec 11 16:54:08 crc kubenswrapper[5109]: --enable-interconnect \ Dec 11 16:54:08 crc kubenswrapper[5109]: --init-cluster-manager "${K8S_NODE}" \ Dec 11 16:54:08 crc kubenswrapper[5109]: --config-file=/run/ovnkube-config/ovnkube.conf \ Dec 11 16:54:08 crc kubenswrapper[5109]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Dec 11 16:54:08 crc kubenswrapper[5109]: --metrics-bind-address "127.0.0.1:29108" \ Dec 11 16:54:08 crc kubenswrapper[5109]: --metrics-enable-pprof \ Dec 11 16:54:08 crc kubenswrapper[5109]: --metrics-enable-config-duration \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${ovn_v4_join_subnet_opt} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${ovn_v6_join_subnet_opt} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${ovn_v4_transit_switch_subnet_opt} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${ovn_v6_transit_switch_subnet_opt} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${dns_name_resolver_enabled_flag} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${persistent_ips_enabled_flag} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${multi_network_enabled_flag} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${network_segmentation_enabled_flag} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${gateway_mode_flags} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${route_advertisements_enable_flag} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${preconfigured_udn_addresses_enable_flag} \ Dec 11 16:54:08 crc kubenswrapper[5109]: --enable-egress-ip=true \ Dec 11 16:54:08 crc kubenswrapper[5109]: --enable-egress-firewall=true \ Dec 11 16:54:08 crc kubenswrapper[5109]: --enable-egress-qos=true \ Dec 11 16:54:08 crc kubenswrapper[5109]: --enable-egress-service=true \ Dec 11 16:54:08 crc kubenswrapper[5109]: --enable-multicast \ Dec 11 16:54:08 crc kubenswrapper[5109]: --enable-multi-external-gateway=true \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${multi_network_policy_enabled_flag} \ Dec 11 16:54:08 crc kubenswrapper[5109]: ${admin_network_policy_enabled_flag} Dec 11 16:54:08 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c4fpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-57b78d8988-dzv7n_openshift-ovn-kubernetes(5f6bde63-48d8-4726-b7d6-0f36f7cf47f9): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:08 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.284376 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hx2kb" event={"ID":"7c45b4ae-02aa-4614-977b-544a9e303bdf","Type":"ContainerStarted","Data":"419ecb6687f142b181cf0a23840a7b085438301c10b68baf51ac446dd3b597e1"} Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.284779 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-5jnd7" podUID="428b39f5-eb1c-4f65-b7a4-eeb6e84860cc" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.284829 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.285368 5109 kuberuntime_manager.go:1358] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l6lpl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-c2wzl_openshift-multus(189d054c-6301-4cf4-ba60-da0380233d01): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.286623 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xqdz5" event={"ID":"61fc568b-d7c2-4633-b28c-f65428d01030","Type":"ContainerStarted","Data":"ae314459752c250a056ca70b989d3c998b65d083f7a27c031e2fbada86a1beff"} Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.286945 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" podUID="189d054c-6301-4cf4-ba60-da0380233d01" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.287678 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:08 crc kubenswrapper[5109]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Dec 11 16:54:08 crc kubenswrapper[5109]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Dec 11 16:54:08 crc kubenswrapper[5109]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nbrkp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-hx2kb_openshift-multus(7c45b4ae-02aa-4614-977b-544a9e303bdf): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:08 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.289194 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:08 crc kubenswrapper[5109]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Dec 11 16:54:08 crc kubenswrapper[5109]: while [ true ]; Dec 11 16:54:08 crc kubenswrapper[5109]: do Dec 11 16:54:08 crc kubenswrapper[5109]: for f in $(ls /tmp/serviceca); do Dec 11 16:54:08 crc kubenswrapper[5109]: echo $f Dec 11 16:54:08 crc kubenswrapper[5109]: ca_file_path="/tmp/serviceca/${f}" Dec 11 16:54:08 crc kubenswrapper[5109]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Dec 11 16:54:08 crc kubenswrapper[5109]: reg_dir_path="/etc/docker/certs.d/${f}" Dec 11 16:54:08 crc kubenswrapper[5109]: if [ -e "${reg_dir_path}" ]; then Dec 11 16:54:08 crc kubenswrapper[5109]: cp -u $ca_file_path $reg_dir_path/ca.crt Dec 11 16:54:08 crc kubenswrapper[5109]: else Dec 11 16:54:08 crc kubenswrapper[5109]: mkdir $reg_dir_path Dec 11 16:54:08 crc kubenswrapper[5109]: cp $ca_file_path $reg_dir_path/ca.crt Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: for d in $(ls /etc/docker/certs.d); do Dec 11 16:54:08 crc kubenswrapper[5109]: echo $d Dec 11 16:54:08 crc kubenswrapper[5109]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Dec 11 16:54:08 crc kubenswrapper[5109]: reg_conf_path="/tmp/serviceca/${dp}" Dec 11 16:54:08 crc kubenswrapper[5109]: if [ ! -e "${reg_conf_path}" ]; then Dec 11 16:54:08 crc kubenswrapper[5109]: rm -rf /etc/docker/certs.d/$d Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: sleep 60 & wait ${!} Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-p72wg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-xqdz5_openshift-image-registry(61fc568b-d7c2-4633-b28c-f65428d01030): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:08 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.289248 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-hx2kb" podUID="7c45b4ae-02aa-4614-977b-544a9e303bdf" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.289423 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rnsbf" event={"ID":"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45","Type":"ContainerStarted","Data":"afac7c21d9f5b75a22cbfc903eaebec8655f4fb4c113b38dfa91bd6c347cc01b"} Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.290406 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-xqdz5" podUID="61fc568b-d7c2-4633-b28c-f65428d01030" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.291467 5109 kuberuntime_manager.go:1358] "Unhandled Error" err=< Dec 11 16:54:08 crc kubenswrapper[5109]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e,Command:[/bin/bash -c #!/bin/bash Dec 11 16:54:08 crc kubenswrapper[5109]: set -uo pipefail Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Dec 11 16:54:08 crc kubenswrapper[5109]: HOSTS_FILE="/etc/hosts" Dec 11 16:54:08 crc kubenswrapper[5109]: TEMP_FILE="/tmp/hosts.tmp" Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: IFS=', ' read -r -a services <<< "${SERVICES}" Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: # Make a temporary file with the old hosts file's attributes. Dec 11 16:54:08 crc kubenswrapper[5109]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Dec 11 16:54:08 crc kubenswrapper[5109]: echo "Failed to preserve hosts file. Exiting." Dec 11 16:54:08 crc kubenswrapper[5109]: exit 1 Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: while true; do Dec 11 16:54:08 crc kubenswrapper[5109]: declare -A svc_ips Dec 11 16:54:08 crc kubenswrapper[5109]: for svc in "${services[@]}"; do Dec 11 16:54:08 crc kubenswrapper[5109]: # Fetch service IP from cluster dns if present. We make several tries Dec 11 16:54:08 crc kubenswrapper[5109]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Dec 11 16:54:08 crc kubenswrapper[5109]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Dec 11 16:54:08 crc kubenswrapper[5109]: # support UDP loadbalancers and require reaching DNS through TCP. Dec 11 16:54:08 crc kubenswrapper[5109]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Dec 11 16:54:08 crc kubenswrapper[5109]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Dec 11 16:54:08 crc kubenswrapper[5109]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Dec 11 16:54:08 crc kubenswrapper[5109]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Dec 11 16:54:08 crc kubenswrapper[5109]: for i in ${!cmds[*]} Dec 11 16:54:08 crc kubenswrapper[5109]: do Dec 11 16:54:08 crc kubenswrapper[5109]: ips=($(eval "${cmds[i]}")) Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: svc_ips["${svc}"]="${ips[@]}" Dec 11 16:54:08 crc kubenswrapper[5109]: break Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: # Update /etc/hosts only if we get valid service IPs Dec 11 16:54:08 crc kubenswrapper[5109]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Dec 11 16:54:08 crc kubenswrapper[5109]: # Stale entries could exist in /etc/hosts if the service is deleted Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ -n "${svc_ips[*]-}" ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Dec 11 16:54:08 crc kubenswrapper[5109]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Dec 11 16:54:08 crc kubenswrapper[5109]: # Only continue rebuilding the hosts entries if its original content is preserved Dec 11 16:54:08 crc kubenswrapper[5109]: sleep 60 & wait Dec 11 16:54:08 crc kubenswrapper[5109]: continue Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: # Append resolver entries for services Dec 11 16:54:08 crc kubenswrapper[5109]: rc=0 Dec 11 16:54:08 crc kubenswrapper[5109]: for svc in "${!svc_ips[@]}"; do Dec 11 16:54:08 crc kubenswrapper[5109]: for ip in ${svc_ips[${svc}]}; do Dec 11 16:54:08 crc kubenswrapper[5109]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: if [[ $rc -ne 0 ]]; then Dec 11 16:54:08 crc kubenswrapper[5109]: sleep 60 & wait Dec 11 16:54:08 crc kubenswrapper[5109]: continue Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: Dec 11 16:54:08 crc kubenswrapper[5109]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Dec 11 16:54:08 crc kubenswrapper[5109]: # Replace /etc/hosts with our modified version if needed Dec 11 16:54:08 crc kubenswrapper[5109]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Dec 11 16:54:08 crc kubenswrapper[5109]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Dec 11 16:54:08 crc kubenswrapper[5109]: fi Dec 11 16:54:08 crc kubenswrapper[5109]: sleep 60 & wait Dec 11 16:54:08 crc kubenswrapper[5109]: unset svc_ips Dec 11 16:54:08 crc kubenswrapper[5109]: done Dec 11 16:54:08 crc kubenswrapper[5109]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-dir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6v94n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-rnsbf_openshift-dns(305d0e9b-83ad-424c-a5a5-c4ba12d3ae45): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Dec 11 16:54:08 crc kubenswrapper[5109]: > logger="UnhandledError" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.292697 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-rnsbf" podUID="305d0e9b-83ad-424c-a5a5-c4ba12d3ae45" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.293083 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-hx2kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c45b4ae-02aa-4614-977b-544a9e303bdf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbrkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hx2kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.318315 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561da642-b358-4ddc-86ee-669164d7b3c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wfxmd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.325280 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.325336 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.325356 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.325382 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.325402 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.332230 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-dzv7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.356676 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d83f88-292e-48c6-a9d2-1613e83851e6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://0ec3b28b6104ac258b2e97c9f10ec3d55ea93bbf9d9165fdce21648201c85e9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://352c8d599c93e387e36a5488dcc02f39161e4d8c0a9fbf050e4e19c3c1a4fd92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://445fc06154e726c5f8c28bf3f36d0bc02dde5ccc34cfefcb928cbee1eb9c0d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ca35d4089afda84920619ba9d77e28fdb8bf432b26bf3cb949151fce931db7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://a3b479ab0d924d9fbb549995801bad80a45ec723aff9b99d07b5d67b37675cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.371120 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c5b2f1d-3171-42d0-8861-bdb91323d73b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T16:53:38Z\\\",\\\"message\\\":\\\"vvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nW1211 16:53:37.766372 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1211 16:53:37.766542 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI1211 16:53:37.769894 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1407815461/tls.crt::/tmp/serving-cert-1407815461/tls.key\\\\\\\"\\\\nI1211 16:53:38.166632 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 16:53:38.168550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 16:53:38.168573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 16:53:38.168605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 16:53:38.168612 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 16:53:38.173779 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 16:53:38.173817 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173826 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 16:53:38.173842 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 16:53:38.173848 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 16:53:38.173853 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 16:53:38.173934 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 16:53:38.175696 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T16:53:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.385301 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.395432 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.407116 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.414898 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnsbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v94n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnsbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.427502 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.427612 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.427690 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.427764 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.427820 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.436330 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"189d054c-6301-4cf4-ba60-da0380233d01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2wzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.445296 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.453227 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad4ed271-a1a8-408d-b043-99cf044fb573\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fp2cw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.464168 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25a760d4-55be-452a-b344-fe5e4ae5b5a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c10e7a3d56908a1837ca1b29a944fbe9febec4030a733cd53c0ecda0177091c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b0e470f587ad2b1f84cb1100ce6155dbe9ba3e4a36c01970e843d5cdf38b3d56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c37b01eb1234360d043eff79f1eb374f5d95c1132e416be3c05ad3fd9fdfc36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.477198 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6525f27e-56ac-4b08-ac25-b854b4edf7fc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6ab4918fcfdbfaadb2441c5d160818ac2fd2f2d718e73fc0ef7d9d62dc004b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.485428 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1f09ed1-77b5-404d-abd1-44afa5fab784\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xgkvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.493602 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xqdz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fc568b-d7c2-4633-b28c-f65428d01030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p72wg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xqdz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.506282 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13d2072-a409-4658-aafa-a35c7c947e4e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://b9608c14d8ea51cb0920addf885ee63ff3e4c8a81ab46a226edebd9a0bb6a399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://58b2adf4e750dc18bc35c0aea18178cf6daed481b29ea518459defa47769e5d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://1cbeee4ade3904eb92ab061e705d6387d66e4f501d3ee96e67260accf10d2acb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.514613 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.514665 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.514693 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.514722 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.514797 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.514817 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.514827 5109 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.514869 5109 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.514875 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:10.514860371 +0000 UTC m=+88.194551837 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.514932 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:10.514917973 +0000 UTC m=+88.194609449 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.515010 5109 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.515027 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.515091 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.515113 5109 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.515127 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:10.515104267 +0000 UTC m=+88.194795743 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.515282 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:10.515211779 +0000 UTC m=+88.194903285 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.518150 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.533788 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.534342 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.534392 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.534414 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.534435 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.534450 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.544180 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad4ed271-a1a8-408d-b043-99cf044fb573\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fp2cw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.558890 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25a760d4-55be-452a-b344-fe5e4ae5b5a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c10e7a3d56908a1837ca1b29a944fbe9febec4030a733cd53c0ecda0177091c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b0e470f587ad2b1f84cb1100ce6155dbe9ba3e4a36c01970e843d5cdf38b3d56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c37b01eb1234360d043eff79f1eb374f5d95c1132e416be3c05ad3fd9fdfc36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.601152 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6525f27e-56ac-4b08-ac25-b854b4edf7fc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6ab4918fcfdbfaadb2441c5d160818ac2fd2f2d718e73fc0ef7d9d62dc004b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.636880 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.636937 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.636953 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.636974 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.636990 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.639135 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1f09ed1-77b5-404d-abd1-44afa5fab784\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xgkvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.681493 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xqdz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fc568b-d7c2-4633-b28c-f65428d01030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p72wg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xqdz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.727149 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13d2072-a409-4658-aafa-a35c7c947e4e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://b9608c14d8ea51cb0920addf885ee63ff3e4c8a81ab46a226edebd9a0bb6a399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://58b2adf4e750dc18bc35c0aea18178cf6daed481b29ea518459defa47769e5d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://1cbeee4ade3904eb92ab061e705d6387d66e4f501d3ee96e67260accf10d2acb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.741039 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.741072 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.741083 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.741107 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.741117 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.764396 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.799376 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.840220 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-hx2kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c45b4ae-02aa-4614-977b-544a9e303bdf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbrkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hx2kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.843997 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.844078 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.844092 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.844137 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.844153 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.886560 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561da642-b358-4ddc-86ee-669164d7b3c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wfxmd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.899911 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.900029 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.900084 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.900262 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.900363 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.900437 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.900455 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.900572 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.907909 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01080b46-74f1-4191-8755-5152a57b3b25" path="/var/lib/kubelet/pods/01080b46-74f1-4191-8755-5152a57b3b25/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.909551 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09cfa50b-4138-4585-a53e-64dd3ab73335" path="/var/lib/kubelet/pods/09cfa50b-4138-4585-a53e-64dd3ab73335/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.912597 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dd0fbac-8c0d-4228-8faa-abbeedabf7db" path="/var/lib/kubelet/pods/0dd0fbac-8c0d-4228-8faa-abbeedabf7db/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.914957 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0effdbcf-dd7d-404d-9d48-77536d665a5d" path="/var/lib/kubelet/pods/0effdbcf-dd7d-404d-9d48-77536d665a5d/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.919202 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="149b3c48-e17c-4a66-a835-d86dabf6ff13" path="/var/lib/kubelet/pods/149b3c48-e17c-4a66-a835-d86dabf6ff13/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.920139 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.920343 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:10.920322089 +0000 UTC m=+88.600013565 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.920507 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.920725 5109 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: E1211 16:54:08.920853 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs podName:ad4ed271-a1a8-408d-b043-99cf044fb573 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:10.920830571 +0000 UTC m=+88.600522077 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs") pod "network-metrics-daemon-fp2cw" (UID: "ad4ed271-a1a8-408d-b043-99cf044fb573") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.922096 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16bdd140-dce1-464c-ab47-dd5798d1d256" path="/var/lib/kubelet/pods/16bdd140-dce1-464c-ab47-dd5798d1d256/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.922700 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-dzv7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.925159 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18f80adb-c1c3-49ba-8ee4-932c851d3897" path="/var/lib/kubelet/pods/18f80adb-c1c3-49ba-8ee4-932c851d3897/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.926939 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20ce4d18-fe25-4696-ad7c-1bd2d6200a3e" path="/var/lib/kubelet/pods/20ce4d18-fe25-4696-ad7c-1bd2d6200a3e/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.927855 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2325ffef-9d5b-447f-b00e-3efc429acefe" path="/var/lib/kubelet/pods/2325ffef-9d5b-447f-b00e-3efc429acefe/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.929701 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="301e1965-1754-483d-b6cc-bfae7038bbca" path="/var/lib/kubelet/pods/301e1965-1754-483d-b6cc-bfae7038bbca/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.932372 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31fa8943-81cc-4750-a0b7-0fa9ab5af883" path="/var/lib/kubelet/pods/31fa8943-81cc-4750-a0b7-0fa9ab5af883/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.934400 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42a11a02-47e1-488f-b270-2679d3298b0e" path="/var/lib/kubelet/pods/42a11a02-47e1-488f-b270-2679d3298b0e/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.936863 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567683bd-0efc-4f21-b076-e28559628404" path="/var/lib/kubelet/pods/567683bd-0efc-4f21-b076-e28559628404/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.938864 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="584e1f4a-8205-47d7-8efb-3afc6017c4c9" path="/var/lib/kubelet/pods/584e1f4a-8205-47d7-8efb-3afc6017c4c9/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.939818 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="593a3561-7760-45c5-8f91-5aaef7475d0f" path="/var/lib/kubelet/pods/593a3561-7760-45c5-8f91-5aaef7475d0f/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.941659 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ebfebf6-3ecd-458e-943f-bb25b52e2718" path="/var/lib/kubelet/pods/5ebfebf6-3ecd-458e-943f-bb25b52e2718/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.942812 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6077b63e-53a2-4f96-9d56-1ce0324e4913" path="/var/lib/kubelet/pods/6077b63e-53a2-4f96-9d56-1ce0324e4913/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.944851 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca" path="/var/lib/kubelet/pods/6a81eec9-f29e-49a0-a15a-f2f5bd2d95ca/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.946135 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.946174 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.946187 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.946203 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.946216 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:08Z","lastTransitionTime":"2025-12-11T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.946790 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6edfcf45-925b-4eff-b940-95b6fc0b85d4" path="/var/lib/kubelet/pods/6edfcf45-925b-4eff-b940-95b6fc0b85d4/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.948018 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ee8fbd3-1f81-4666-96da-5afc70819f1a" path="/var/lib/kubelet/pods/6ee8fbd3-1f81-4666-96da-5afc70819f1a/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.950370 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a" path="/var/lib/kubelet/pods/71c8ffbe-59c6-4e7d-aa1a-bbd315b3414a/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.954001 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="736c54fe-349c-4bb9-870a-d1c1d1c03831" path="/var/lib/kubelet/pods/736c54fe-349c-4bb9-870a-d1c1d1c03831/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.954977 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7599e0b6-bddf-4def-b7f2-0b32206e8651" path="/var/lib/kubelet/pods/7599e0b6-bddf-4def-b7f2-0b32206e8651/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.956681 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7afa918d-be67-40a6-803c-d3b0ae99d815" path="/var/lib/kubelet/pods/7afa918d-be67-40a6-803c-d3b0ae99d815/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.957658 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df94c10-441d-4386-93a6-6730fb7bcde0" path="/var/lib/kubelet/pods/7df94c10-441d-4386-93a6-6730fb7bcde0/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.958921 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fcc6409-8a0f-44c3-89e7-5aecd7610f8a" path="/var/lib/kubelet/pods/7fcc6409-8a0f-44c3-89e7-5aecd7610f8a/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.959842 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e39f7b-62e4-4fc9-992a-6535ce127a02" path="/var/lib/kubelet/pods/81e39f7b-62e4-4fc9-992a-6535ce127a02/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.960768 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="869851b9-7ffb-4af0-b166-1d8aa40a5f80" path="/var/lib/kubelet/pods/869851b9-7ffb-4af0-b166-1d8aa40a5f80/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.963670 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff" path="/var/lib/kubelet/pods/9276f8f5-2f24-48e1-ab6d-1aab0d8ec3ff/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.964309 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92dfbade-90b6-4169-8c07-72cff7f2c82b" path="/var/lib/kubelet/pods/92dfbade-90b6-4169-8c07-72cff7f2c82b/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.965795 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a6e063-3d1a-4d44-875d-185291448c31" path="/var/lib/kubelet/pods/94a6e063-3d1a-4d44-875d-185291448c31/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.966899 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f71a554-e414-4bc3-96d2-674060397afe" path="/var/lib/kubelet/pods/9f71a554-e414-4bc3-96d2-674060397afe/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.968905 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a208c9c2-333b-4b4a-be0d-bc32ec38a821" path="/var/lib/kubelet/pods/a208c9c2-333b-4b4a-be0d-bc32ec38a821/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.970112 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52afe44-fb37-46ed-a1f8-bf39727a3cbe" path="/var/lib/kubelet/pods/a52afe44-fb37-46ed-a1f8-bf39727a3cbe/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.971538 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a555ff2e-0be6-46d5-897d-863bb92ae2b3" path="/var/lib/kubelet/pods/a555ff2e-0be6-46d5-897d-863bb92ae2b3/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.972402 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7a88189-c967-4640-879e-27665747f20c" path="/var/lib/kubelet/pods/a7a88189-c967-4640-879e-27665747f20c/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.973683 5109 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="af33e427-6803-48c2-a76a-dd9deb7cbf9a" path="/var/lib/kubelet/pods/af33e427-6803-48c2-a76a-dd9deb7cbf9a/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.973831 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af33e427-6803-48c2-a76a-dd9deb7cbf9a" path="/var/lib/kubelet/pods/af33e427-6803-48c2-a76a-dd9deb7cbf9a/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.977183 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af41de71-79cf-4590-bbe9-9e8b848862cb" path="/var/lib/kubelet/pods/af41de71-79cf-4590-bbe9-9e8b848862cb/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.978783 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a" path="/var/lib/kubelet/pods/b05a4c1d-fa93-4d3d-b6e5-235473e1ae2a/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.979462 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d83f88-292e-48c6-a9d2-1613e83851e6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://0ec3b28b6104ac258b2e97c9f10ec3d55ea93bbf9d9165fdce21648201c85e9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://352c8d599c93e387e36a5488dcc02f39161e4d8c0a9fbf050e4e19c3c1a4fd92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://445fc06154e726c5f8c28bf3f36d0bc02dde5ccc34cfefcb928cbee1eb9c0d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ca35d4089afda84920619ba9d77e28fdb8bf432b26bf3cb949151fce931db7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://a3b479ab0d924d9fbb549995801bad80a45ec723aff9b99d07b5d67b37675cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.979944 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4750666-1362-4001-abd0-6f89964cc621" path="/var/lib/kubelet/pods/b4750666-1362-4001-abd0-6f89964cc621/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.981322 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b605f283-6f2e-42da-a838-54421690f7d0" path="/var/lib/kubelet/pods/b605f283-6f2e-42da-a838-54421690f7d0/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.981930 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c491984c-7d4b-44aa-8c1e-d7974424fa47" path="/var/lib/kubelet/pods/c491984c-7d4b-44aa-8c1e-d7974424fa47/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.983606 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5f2bfad-70f6-4185-a3d9-81ce12720767" path="/var/lib/kubelet/pods/c5f2bfad-70f6-4185-a3d9-81ce12720767/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.984555 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc85e424-18b2-4924-920b-bd291a8c4b01" path="/var/lib/kubelet/pods/cc85e424-18b2-4924-920b-bd291a8c4b01/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.985219 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce090a97-9ab6-4c40-a719-64ff2acd9778" path="/var/lib/kubelet/pods/ce090a97-9ab6-4c40-a719-64ff2acd9778/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.986551 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d19cb085-0c5b-4810-b654-ce7923221d90" path="/var/lib/kubelet/pods/d19cb085-0c5b-4810-b654-ce7923221d90/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.987613 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d45be74c-0d98-4d18-90e4-f7ef1b6daaf7" path="/var/lib/kubelet/pods/d45be74c-0d98-4d18-90e4-f7ef1b6daaf7/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.989244 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d565531a-ff86-4608-9d19-767de01ac31b" path="/var/lib/kubelet/pods/d565531a-ff86-4608-9d19-767de01ac31b/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.990092 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7e8f42f-dc0e-424b-bb56-5ec849834888" path="/var/lib/kubelet/pods/d7e8f42f-dc0e-424b-bb56-5ec849834888/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.991341 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9" path="/var/lib/kubelet/pods/dcd10325-9ba5-4a3b-8e4a-e57e3bf210f9/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.992175 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e093be35-bb62-4843-b2e8-094545761610" path="/var/lib/kubelet/pods/e093be35-bb62-4843-b2e8-094545761610/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.993592 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1d2a42d-af1d-4054-9618-ab545e0ed8b7" path="/var/lib/kubelet/pods/e1d2a42d-af1d-4054-9618-ab545e0ed8b7/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.995140 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f559dfa3-3917-43a2-97f6-61ddfda10e93" path="/var/lib/kubelet/pods/f559dfa3-3917-43a2-97f6-61ddfda10e93/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.996619 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65c0ac1-8bca-454d-a2e6-e35cb418beac" path="/var/lib/kubelet/pods/f65c0ac1-8bca-454d-a2e6-e35cb418beac/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.997964 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4" path="/var/lib/kubelet/pods/f7648cbb-48eb-4ba8-87ec-eb096b8fa1e4/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.998922 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7e2c886-118e-43bb-bef1-c78134de392b" path="/var/lib/kubelet/pods/f7e2c886-118e-43bb-bef1-c78134de392b/volumes" Dec 11 16:54:08 crc kubenswrapper[5109]: I1211 16:54:08.999848 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc8db2c7-859d-47b3-a900-2bd0c0b2973b" path="/var/lib/kubelet/pods/fc8db2c7-859d-47b3-a900-2bd0c0b2973b/volumes" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.003330 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c5b2f1d-3171-42d0-8861-bdb91323d73b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T16:53:38Z\\\",\\\"message\\\":\\\"vvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nW1211 16:53:37.766372 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1211 16:53:37.766542 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI1211 16:53:37.769894 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1407815461/tls.crt::/tmp/serving-cert-1407815461/tls.key\\\\\\\"\\\\nI1211 16:53:38.166632 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 16:53:38.168550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 16:53:38.168573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 16:53:38.168605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 16:53:38.168612 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 16:53:38.173779 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 16:53:38.173817 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173826 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 16:53:38.173842 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 16:53:38.173848 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 16:53:38.173853 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 16:53:38.173934 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 16:53:38.175696 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T16:53:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.042069 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.048722 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.048845 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.048868 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.048895 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.048913 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.083328 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.123255 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.152333 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.152418 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.152440 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.152469 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.152487 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.159424 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnsbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v94n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnsbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.208949 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"189d054c-6301-4cf4-ba60-da0380233d01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2wzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.246052 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.263842 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.263920 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.263942 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.263968 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.263987 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.365720 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.365811 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.365829 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.365856 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.365874 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.468416 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.468478 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.468496 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.468521 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.468542 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.570963 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.571030 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.571048 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.571076 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.571095 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.674289 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.674361 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.674386 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.674417 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.674442 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.777170 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.777258 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.777284 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.777318 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.777347 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.879486 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.879538 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.879552 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.879571 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.879585 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.981448 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.981519 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.981532 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.981548 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:09 crc kubenswrapper[5109]: I1211 16:54:09.981560 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:09Z","lastTransitionTime":"2025-12-11T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.084908 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.084969 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.084987 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.085010 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.085028 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.188229 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.188295 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.188315 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.188340 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.188358 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.291465 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.291559 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.291585 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.291616 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.291636 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.394221 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.394285 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.394296 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.394326 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.394335 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.497129 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.497256 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.497277 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.497301 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.497319 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.541973 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.542018 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.542041 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.542066 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542101 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542123 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542135 5109 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542166 5109 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542174 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542193 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542203 5109 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542194 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:14.54217523 +0000 UTC m=+92.221866696 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542245 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:14.542234951 +0000 UTC m=+92.221926417 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542259 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:14.542252012 +0000 UTC m=+92.221943478 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542269 5109 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.542294 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:14.542286043 +0000 UTC m=+92.221977509 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.601233 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.601340 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.601370 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.601408 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.601444 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.704636 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.704692 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.704711 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.704734 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.704792 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.807814 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.808391 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.808595 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.808874 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.809078 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.899340 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.899523 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.899577 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.899644 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.899776 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.900062 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.900334 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.900047 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.912046 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.912336 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.912529 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.912778 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.913001 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:10Z","lastTransitionTime":"2025-12-11T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.947260 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.947525 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:14.947486304 +0000 UTC m=+92.627177830 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:10 crc kubenswrapper[5109]: I1211 16:54:10.947899 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.948165 5109 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:10 crc kubenswrapper[5109]: E1211 16:54:10.948278 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs podName:ad4ed271-a1a8-408d-b043-99cf044fb573 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:14.948253743 +0000 UTC m=+92.627945249 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs") pod "network-metrics-daemon-fp2cw" (UID: "ad4ed271-a1a8-408d-b043-99cf044fb573") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.015370 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.015442 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.015470 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.015500 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.015522 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.118812 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.118907 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.118935 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.118962 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.118981 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.221855 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.221922 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.221943 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.221975 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.221997 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.324809 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.324880 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.324906 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.324942 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.324965 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.428115 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.428178 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.428196 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.428223 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.428241 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.531565 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.531643 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.531664 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.531695 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.531718 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.634355 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.634427 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.634446 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.634474 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.634493 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.648629 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.648695 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.648713 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.648760 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.648783 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: E1211 16:54:11.664238 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.669645 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.669705 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.669724 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.669785 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.669810 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: E1211 16:54:11.684616 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.689700 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.689952 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.690104 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.690249 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.690395 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: E1211 16:54:11.708858 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.713548 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.713615 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.713642 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.713674 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.713698 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: E1211 16:54:11.727571 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.732034 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.732125 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.732148 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.732177 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.732201 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: E1211 16:54:11.745381 5109 kubelet_node_status.go:597] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32400460Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32861260Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c8a088031661d94022418e93fb63744c38e1c4cff93ea3b95c096a290c2b7a3\\\"],\\\"sizeBytes\\\":2981840865},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\"],\\\"sizeBytes\\\":1641503854},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:286bb0beab328954b0a86b7f066fd5a843b462d6acb2812df7ec788015cd32d4\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:be02784ed82978c399102be1c6c9f2ca441be4d984e0fd7100c155dd4417ebbf\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1597684406},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\"],\\\"sizeBytes\\\":1261384762},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:0d50962980a5aeecae2d99c98913fb0f46940164e41de0af2ba0e3dafe0d9017\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:8d607fb6cc75ca36bca1e0a9c5bea5d1919b75db20733df69c64c8a10ee8083d\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1224304325},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:541db5b20a3d2199602b3b5ac80f09ea31498034e9ae3841238b03a39150f0d7\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a4c5df55584cba56f00004a090923a5c6de2071add5eb1672a5e20aa646aad8c\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.20\\\"],\\\"sizeBytes\\\":1126957757},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:b1c859067d6b7b785ab4977ed7137c5b3bb257234f7d7737a1d2836cef1576b5\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:df08951924aa23b2333436a1d04b2dba56c366bb4f09d39ae3aedb980e4fb909\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.20\\\"],\\\"sizeBytes\\\":1079537324},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\"],\\\"sizeBytes\\\":1052707833},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8d1a1e4abe0326c3af89e9eaa4b7449dd2d5b6f9403c677e19b00b24947b1df9\\\"],\\\"sizeBytes\\\":989392005},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b2b1fc3d5bb4944cbd5b23b87566d7ba24b1b66f5a0465f76bcc05023191cc47\\\"],\\\"sizeBytes\\\":971668163},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\"],\\\"sizeBytes\\\":969078739},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3d6c8802ae53d6aecf38aa7b560d7892193806bdeb3d7c1637fac77c47fd1f\\\"],\\\"sizeBytes\\\":876488654},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\"],\\\"sizeBytes\\\":847332502},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:36c4867005702f0c4cbfcfa33f18a98596a6c9b1340b633c85ccef84a0c4f889\\\"],\\\"sizeBytes\\\":769516783},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\"],\\\"sizeBytes\\\":721591926},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\"],\\\"sizeBytes\\\":646867625},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\"],\\\"sizeBytes\\\":638910445},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae245c97fc463e876c3024efb806fa8f4efb13b3f06f1bdd3e7e1447f5a5dce4\\\"],\\\"sizeBytes\\\":617699779},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d4926e304011637ca9df370a193896d685f0f3ffabbec234ec827abdbeb083f9\\\"],\\\"sizeBytes\\\":607756695},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\"],\\\"sizeBytes\\\":584721741},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\"],\\\"sizeBytes\\\":545674969},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:574d49b89604b8e8103abf57feee77812fe8cf441eafc17fdff95d57ca80645e\\\"],\\\"sizeBytes\\\":542463064},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\"],\\\"sizeBytes\\\":539380592},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9506bdcf97d5200cf2cf4cdf110aebafdd141a24f6589bf1e1cfe27bb7fc1ed2\\\"],\\\"sizeBytes\\\":533027808},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\"],\\\"sizeBytes\\\":528200501},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\"],\\\"sizeBytes\\\":527774342},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e5e8108294b086fdb797365e5a46badba9b3d866bdcddc8460a51e05a253753d\\\"],\\\"sizeBytes\\\":526632426},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5827f6ae3beb4853192e02cc18890467bd251b33070f36f9a105991e7e6d3c9b\\\"],\\\"sizeBytes\\\":522490210},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:66c8fe5d45ff249643dae75185dd2787ea1b0ae87d5699a8222149c07689557c\\\"],\\\"sizeBytes\\\":520141094},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:baf975b6944f2844860c440636e0d4b80b2fdc473d30f32ae7d6989f2fc2b135\\\"],\\\"sizeBytes\\\":519815758},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:765cf9c3ebf4df049ebc022beaaf52f52852cf89fb802034536ad91dd45db807\\\"],\\\"sizeBytes\\\":519539350},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:52e442bc8198ac925caff87ddd35b3107b7375d5afc9c2eb041ca4e79db72c6f\\\"],\\\"sizeBytes\\\":518690683},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:43b0e0b7e1955ee905e48799a62f50b8a8df553190415ce1f5550375c2507ca5\\\"],\\\"sizeBytes\\\":518251952},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:977a316fa3598eb575a4477dafc09bbf06fad21c4ec2867052225d74f2a9f366\\\"],\\\"sizeBytes\\\":511136541},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\"],\\\"sizeBytes\\\":510122097},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dbd8603d717c26901bcf9731b1e0392ae4bc08a270ed1eeb45839e44bed9607d\\\"],\\\"sizeBytes\\\":508941917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\"],\\\"sizeBytes\\\":508318343},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7a726c68cebc9b08edd734a8bae5150ae5950f7734fe9b9c2a6e0d06f21cc095\\\"],\\\"sizeBytes\\\":498380948},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:82501261b9c63012ba3b83fe4d6703c0af5eb9c9151670eb90ae480b9507d761\\\"],\\\"sizeBytes\\\":497232440},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:4e4239621caed0b0d9132d167403631e9af86be9a395977f013e201ead281bb4\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:c0b1bec73fdb6853eb3bd9e9733aee2d760ca09a33cfd94adf9ab7b706e83fa9\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":491224335},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b0f7abf2f97afd1127d9245d764338c6047bac1711b2cee43112570a85946360\\\"],\\\"sizeBytes\\\":490381192},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:21b12ff0c81c1d535e7c31aff3a73b1e9ca763e5f88037f59ade0dfab6ed8946\\\"],\\\"sizeBytes\\\":482632652},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:036ed6efe4cb5f5b90ee7f9ef5297c8591b8d67aa36b3c58b4fc5417622a140c\\\"],\\\"sizeBytes\\\":477561861},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0fe5a041a2b99d736e82f1b4a6cd9792c5e23ded475e9f0742cd19234070f989\\\"],\\\"sizeBytes\\\":475327956},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\"],\\\"sizeBytes\\\":475137830},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2599f32933f5fea6066ede54ad8f6150adb7bd9067892f251d5913121d5c630d\\\"],\\\"sizeBytes\\\":472771950},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:651bbe9d418f49c2c889d731df67cf5d88dff59dc03f5a1b5d4c8bb3ae001f1a\\\"],\\\"sizeBytes\\\":469976318},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:4fe612a1572df462d6a4b664a10bc2e6cad239648acbf8c0303f8fca5d2596c0\\\"],\\\"sizeBytes\\\":468393024},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a5bb05344dd2296077f5066e908ede0eea23f5a12fb78ef86a9513c88d3faaca\\\"],\\\"sizeBytes\\\":464375011},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\"],\\\"sizeBytes\\\":462844959}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2bb2c55e-c8c8-470b-a3e3-d0e526c54090\\\",\\\"systemUUID\\\":\\\"789f6dfb-7431-471d-a595-070518eb3889\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:11 crc kubenswrapper[5109]: E1211 16:54:11.745778 5109 kubelet_node_status.go:584] "Unable to update node status" err="update node status exceeds retry count" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.747674 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.747942 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.748105 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.748304 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.748467 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.851671 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.851785 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.851806 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.851832 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.851854 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.954069 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.954128 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.954147 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.954171 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:11 crc kubenswrapper[5109]: I1211 16:54:11.954190 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:11Z","lastTransitionTime":"2025-12-11T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.057072 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.057133 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.057160 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.057198 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.057224 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.160146 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.160219 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.160244 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.160280 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.160306 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.263000 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.263374 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.263550 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.263689 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.263873 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.366233 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.366294 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.366311 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.366334 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.366352 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.468918 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.469000 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.469023 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.469048 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.469066 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.571895 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.572638 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.572804 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.572943 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.573056 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.675817 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.675915 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.675935 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.675960 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.675978 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.778011 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.778086 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.778106 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.778128 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.778147 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.880465 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.880527 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.880545 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.880569 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.880588 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.899932 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:12 crc kubenswrapper[5109]: E1211 16:54:12.900145 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.900184 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.900282 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:12 crc kubenswrapper[5109]: E1211 16:54:12.900373 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.900165 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:12 crc kubenswrapper[5109]: E1211 16:54:12.900709 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:12 crc kubenswrapper[5109]: E1211 16:54:12.900909 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.920104 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b13d2072-a409-4658-aafa-a35c7c947e4e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://b9608c14d8ea51cb0920addf885ee63ff3e4c8a81ab46a226edebd9a0bb6a399\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c6a47106effd9e9a41131e2bf6c832b80cd77b3439334f760b35b0729f2fb00\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"60m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://58b2adf4e750dc18bc35c0aea18178cf6daed481b29ea518459defa47769e5d0\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://1cbeee4ade3904eb92ab061e705d6387d66e4f501d3ee96e67260accf10d2acb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:f69b9cc9b9cfde726109a9e12b80a3eefa472d7e29159df0fbc7143c48983cd6\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-trust-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/kubernetes\\\",\\\"name\\\":\\\"var-run-kubernetes\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.937922 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f863fff9-286a-45fa-b8f0-8a86994b8440\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l7w75\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-5bb8f5cd97-xdvz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.960462 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fc4541ce-7789-4670-bc75-5c2868e52ce0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8nt2j\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-dgvkt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.978668 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-hx2kb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7c45b4ae-02aa-4614-977b-544a9e303bdf\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-nbrkp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-hx2kb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.984428 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.984680 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.985240 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.985448 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:12 crc kubenswrapper[5109]: I1211 16:54:12.985605 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:12Z","lastTransitionTime":"2025-12-11T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.007273 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"561da642-b358-4ddc-86ee-669164d7b3c5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-46pxq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-wfxmd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.022482 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174629230f874ae7d9ceda909ef45aced0cc8b21537851a0aceca55b0685b122\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-c4fpt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-57b78d8988-dzv7n\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.053105 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9d83f88-292e-48c6-a9d2-1613e83851e6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"},\\\"containerID\\\":\\\"cri-o://0ec3b28b6104ac258b2e97c9f10ec3d55ea93bbf9d9165fdce21648201c85e9c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"300m\\\",\\\"memory\\\":\\\"600Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"},\\\"containerID\\\":\\\"cri-o://352c8d599c93e387e36a5488dcc02f39161e4d8c0a9fbf050e4e19c3c1a4fd92\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"40m\\\",\\\"memory\\\":\\\"200Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://445fc06154e726c5f8c28bf3f36d0bc02dde5ccc34cfefcb928cbee1eb9c0d85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://ca35d4089afda84920619ba9d77e28fdb8bf432b26bf3cb949151fce931db7a1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:bf05b9b2ba66351a6c59f4259fb377f62237a00af3b4f0b95f64409e2f25770e\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:48Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://a3b479ab0d924d9fbb549995801bad80a45ec723aff9b99d07b5d67b37675cc6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:47Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c13bdf057bfb62169dd0639427dbd6cfeec0563a87cbeba87bee8cee4aa553cd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd-auto-backup\\\",\\\"name\\\":\\\"etcd-auto-backup-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ee8ee16c21db95f89957afa8063b33b99818126b178011f541b7ec7934d1caf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"},\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9e388ee2b3562b6267447cbcc4b95ca7a61bf361840d36a682480da671b83612\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"60Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://35c7f9106fc2a930fac553475dff76b86053a6e8b264f1b185c065b7aa8864cf\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.074821 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4c5b2f1d-3171-42d0-8861-bdb91323d73b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"},\\\"containerID\\\":\\\"cri-o://680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"265m\\\",\\\"memory\\\":\\\"1Gi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/etc/pki/ca-trust/extracted/pem\\\",\\\"name\\\":\\\"ca-bundle-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-11T16:53:38Z\\\",\\\"message\\\":\\\"vvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nW1211 16:53:37.766372 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI1211 16:53:37.766542 1 builder.go:304] check-endpoints version v0.0.0-unknown-c3d9642-c3d9642\\\\nI1211 16:53:37.769894 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1407815461/tls.crt::/tmp/serving-cert-1407815461/tls.key\\\\\\\"\\\\nI1211 16:53:38.166632 1 requestheader_controller.go:255] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1211 16:53:38.168550 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1211 16:53:38.168573 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1211 16:53:38.168605 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1211 16:53:38.168612 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1211 16:53:38.173779 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1211 16:53:38.173817 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173826 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1211 16:53:38.173835 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1211 16:53:38.173842 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1211 16:53:38.173848 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1211 16:53:38.173853 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1211 16:53:38.173934 1 genericapiserver.go:546] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1211 16:53:38.175696 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-11T16:53:37Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"10m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(3a14caf222afb62aaabdc47808b6f944)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:46Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.088620 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"17b87002-b798-480a-8e17-83053d698239\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gwt8b\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-fhkjl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.088946 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.088992 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.089006 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.089022 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.089035 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.104758 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dsgwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-5jnd7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.120312 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fbdfe828b092b23e6d4480daf3e0216aada6debaf1ef1b314a0a31e73ebf13c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-5ff7774fd9-nljh6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.134236 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-dns/node-resolver-rnsbf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5c5d7468f6838b6a714482e62ea956659212f3415ec8f69989f75eb6d8744a6e\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6v94n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-rnsbf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.147166 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"189d054c-6301-4cf4-ba60-da0380233d01\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:85f1323d589d7af13b096b1f9b438b9dfe08f3fab37534e2780e6490a665bf05\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5a2a7b3c2f1598189d8880e6aa15ab11a65b201f25012f77ba41e7487a60729a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1b55c029f731ebbde3c5580eef98a588264f4d6a8ae667805c9521dd1ecf1d5d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6efa070ceb93cc5fc2e76eab6d9c96ac3c4f8812085d0b6eb6e3f513b5bac782\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3454e762466e22e2a893650b9781823558bc6fdfda2aa4188aff3cb819014c4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:be2edaed22535093bdb486afe5960ff4f3b0bd96f88dc1753b584cc28184a0b0\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/etc/whereabouts/config\\\",\\\"name\\\":\\\"whereabouts-flatfile-configmap\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l6lpl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2wzl\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.165198 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34177974-8d82-49d2-a763-391d0df3bbd8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:06Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a8581a82ba5c8343a743aa302c4848249d8c32a9f2cd10fa68d89d835a1bdf8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7xz2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-7bdcf4f5bd-7fjxv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.177503 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ad4ed271-a1a8-408d-b043-99cf044fb573\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:49b34ce0d25eec7a6077f4bf21bf7d4e64e598d28785a20b9ee3594423b7de14\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-x5rg8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-fp2cw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.192129 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.192451 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.192580 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.192727 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.193011 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.193209 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"25a760d4-55be-452a-b344-fe5e4ae5b5a0\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:53:30Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c10e7a3d56908a1837ca1b29a944fbe9febec4030a733cd53c0ecda0177091c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://b0e470f587ad2b1f84cb1100ce6155dbe9ba3e4a36c01970e843d5cdf38b3d56\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://c37b01eb1234360d043eff79f1eb374f5d95c1132e416be3c05ad3fd9fdfc36c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e504172345491d90bbbf1e7e45488e73073f4c6d7c2355245871051596fc85db\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp\\\",\\\"name\\\":\\\"tmp\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8a46fa8feeea5d04fd602559027f8bacc97e12bbf8e33793dca08e812e1f8825\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"15m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f7f14166126f1a8367f3298ce7ae55e5a0d39ff0bb26e752b0f3f73a74d15ae3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":0,\\\"supplementalGroups\\\":[0],\\\"uid\\\":0}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.205582 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6525f27e-56ac-4b08-ac25-b854b4edf7fc\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:52:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://6ab4918fcfdbfaadb2441c5d160818ac2fd2f2d718e73fc0ef7d9d62dc004b11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"20m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-11T16:52:45Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"allocatedResources\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"},\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"resources\\\":{\\\"requests\\\":{\\\"cpu\\\":\\\"5m\\\",\\\"memory\\\":\\\"50Mi\\\"}},\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e51e131075fd3f4e55c394837bdb8cbace7b8eda9e5056b5715ef0ba45b82014\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-11T16:52:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-11T16:52:44Z\\\"}},\\\"user\\\":{\\\"linux\\\":{\\\"gid\\\":65534,\\\"supplementalGroups\\\":[65534],\\\"uid\\\":65534}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:52:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.218580 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1f09ed1-77b5-404d-abd1-44afa5fab784\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:16d5a229c172bde2f4238e8a88602fd6351d80b262f35484740a979d8b3567a5\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9414357f9345a841e0565265700ecc6637f846c83bd5908dbb7b306432465115\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5vmt7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-xgkvv\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.232635 5109 status_manager.go:919] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xqdz5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"61fc568b-d7c2-4633-b28c-f65428d01030\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-11T16:54:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dcb03ccba25366bbdf74cbab6738e7ef1f97f62760886ec445a40cdf29b60418\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-p72wg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-11T16:54:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xqdz5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.295294 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.295370 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.295386 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.295403 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.295415 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.397492 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.397561 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.397585 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.397617 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.397641 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.500072 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.500143 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.500163 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.500189 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.500209 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.602145 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.602189 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.602201 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.602219 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.602231 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.704068 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.704111 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.704120 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.704134 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.704145 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.806549 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.806590 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.806619 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.806633 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.806642 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.908583 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.908645 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.908663 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.908681 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:13 crc kubenswrapper[5109]: I1211 16:54:13.908696 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:13Z","lastTransitionTime":"2025-12-11T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.010709 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.010773 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.010785 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.010803 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.010814 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.113007 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.113080 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.113099 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.113127 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.113149 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.215729 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.215861 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.215889 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.215921 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.215952 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.318529 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.318602 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.318625 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.318653 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.318675 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.420839 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.420928 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.420947 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.420973 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.420996 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.524025 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.524110 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.524135 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.524158 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.524175 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.588446 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.588523 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.588567 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.588607 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588666 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588710 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588733 5109 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588823 5109 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588845 5109 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588775 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588921 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:22.588889496 +0000 UTC m=+100.268581002 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588935 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588951 5109 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588960 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:22.588941568 +0000 UTC m=+100.268633074 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.588997 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:22.588975928 +0000 UTC m=+100.268667434 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.589038 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:22.589023689 +0000 UTC m=+100.268715185 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.627214 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.627281 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.627299 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.627321 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.627339 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.729858 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.729921 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.729944 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.729972 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.729990 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.833197 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.833278 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.833302 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.833329 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.833350 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.899978 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.900028 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.900060 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.900172 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.900278 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.900487 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.900615 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.900852 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.935528 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.935713 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.935823 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.935907 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.935966 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:14Z","lastTransitionTime":"2025-12-11T16:54:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.992852 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:14 crc kubenswrapper[5109]: I1211 16:54:14.993123 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.993385 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:22.99335173 +0000 UTC m=+100.673043226 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.993387 5109 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:14 crc kubenswrapper[5109]: E1211 16:54:14.993808 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs podName:ad4ed271-a1a8-408d-b043-99cf044fb573 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:22.99378016 +0000 UTC m=+100.673471666 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs") pod "network-metrics-daemon-fp2cw" (UID: "ad4ed271-a1a8-408d-b043-99cf044fb573") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.038409 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.038468 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.038486 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.038533 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.038551 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.140995 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.141232 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.141451 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.141665 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.141880 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.244149 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.244212 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.244229 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.244253 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.244272 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.347615 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.347692 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.347717 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.347777 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.347797 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.450967 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.451036 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.451060 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.451168 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.451195 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.554503 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.554578 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.554599 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.554623 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.554643 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.657950 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.658004 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.658019 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.658042 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.658058 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.760427 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.760489 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.760514 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.760545 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.760568 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.863126 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.863178 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.863194 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.863218 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.863237 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.967873 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.967955 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.967975 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.968001 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:15 crc kubenswrapper[5109]: I1211 16:54:15.968018 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:15Z","lastTransitionTime":"2025-12-11T16:54:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.069877 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.069920 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.069929 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.069947 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.069957 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.078237 5109 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.172424 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.172498 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.172513 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.172535 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.172549 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.274926 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.274989 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.275009 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.275035 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.275054 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.377851 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.378209 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.378466 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.378710 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.378966 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.481847 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.481920 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.481948 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.481979 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.482002 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.584537 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.584621 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.584635 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.584684 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.584698 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.686904 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.686976 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.687000 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.687032 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.687052 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.789978 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.790066 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.790090 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.790120 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.790145 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.892330 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.892406 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.892430 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.892461 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.892483 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.900075 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:16 crc kubenswrapper[5109]: E1211 16:54:16.900230 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.900821 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:16 crc kubenswrapper[5109]: E1211 16:54:16.900998 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.901116 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.901159 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:16 crc kubenswrapper[5109]: E1211 16:54:16.901293 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:16 crc kubenswrapper[5109]: E1211 16:54:16.901461 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.995003 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.995069 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.995093 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.995124 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:16 crc kubenswrapper[5109]: I1211 16:54:16.995148 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:16Z","lastTransitionTime":"2025-12-11T16:54:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.097891 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.097965 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.097991 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.098022 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.098046 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.200139 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.200225 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.200250 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.200279 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.200302 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.306576 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.306636 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.306663 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.306692 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.306714 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.409390 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.409462 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.409483 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.409511 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.409533 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.511608 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.511687 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.511715 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.511776 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.511799 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.613934 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.613971 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.613980 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.613994 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.614003 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.716387 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.716431 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.716443 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.716456 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.716466 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.819501 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.819866 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.819886 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.819913 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.819936 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.923375 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.923505 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.923531 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.923565 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:17 crc kubenswrapper[5109]: I1211 16:54:17.923589 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:17Z","lastTransitionTime":"2025-12-11T16:54:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.026562 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.026677 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.026698 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.026725 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.026774 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.129323 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.129402 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.129428 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.129458 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.129481 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.231903 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.231960 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.231980 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.232004 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.232022 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.333860 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.333916 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.333934 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.333956 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.333974 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.436108 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.436179 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.436203 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.436233 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.436261 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.538472 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.538523 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.538542 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.538566 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.538584 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.640781 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.640833 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.640850 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.640873 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.640891 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.743145 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.743211 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.743231 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.743258 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.743276 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.845648 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.845724 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.845772 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.845797 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.845815 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.900024 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.900319 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:18 crc kubenswrapper[5109]: E1211 16:54:18.900322 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.900566 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.900662 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:18 crc kubenswrapper[5109]: E1211 16:54:18.900574 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:18 crc kubenswrapper[5109]: E1211 16:54:18.900895 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:18 crc kubenswrapper[5109]: E1211 16:54:18.901127 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.949059 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.949125 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.949143 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.949170 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:18 crc kubenswrapper[5109]: I1211 16:54:18.949189 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:18Z","lastTransitionTime":"2025-12-11T16:54:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.051938 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.052018 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.052044 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.052079 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.052106 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.154419 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.154485 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.154504 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.154543 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.154562 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.257320 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.257364 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.257376 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.257392 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.257403 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.360127 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.360276 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.360302 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.360371 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.360397 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.463063 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.463145 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.463170 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.463201 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.463223 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.565857 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.565919 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.565936 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.565960 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.565978 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.668601 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.668682 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.668708 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.668779 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.668807 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.771650 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.771719 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.771774 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.771802 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.771820 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.874580 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.874662 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.874683 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.874709 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.874727 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.900504 5109 scope.go:117] "RemoveContainer" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.977935 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.977994 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.978012 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.978036 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:19 crc kubenswrapper[5109]: I1211 16:54:19.978055 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:19Z","lastTransitionTime":"2025-12-11T16:54:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.081609 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.081682 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.081702 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.081826 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.081851 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.185524 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.185637 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.185661 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.185722 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.185809 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.288016 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.288281 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.288294 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.288312 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.288324 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.341603 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"d975c83c59994a6c86ff7586f76e25c276b186ade7d5e06e8443018fcd937ab7"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.341663 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-dgvkt" event={"ID":"fc4541ce-7789-4670-bc75-5c2868e52ce0","Type":"ContainerStarted","Data":"352516e3e282de07e48be2ea1e0b6c4b66342b375025018e92c264cad04e48e5"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.383573 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=14.383537048 podStartE2EDuration="14.383537048s" podCreationTimestamp="2025-12-11 16:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:20.383523597 +0000 UTC m=+98.063215063" watchObservedRunningTime="2025-12-11 16:54:20.383537048 +0000 UTC m=+98.063228514" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.392296 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.392373 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.392407 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.392434 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.392449 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.495711 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.495772 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.495783 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.495801 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.495813 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.530074 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=14.530055264 podStartE2EDuration="14.530055264s" podCreationTimestamp="2025-12-11 16:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:20.529789548 +0000 UTC m=+98.209481034" watchObservedRunningTime="2025-12-11 16:54:20.530055264 +0000 UTC m=+98.209746730" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.552432 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=14.552413624 podStartE2EDuration="14.552413624s" podCreationTimestamp="2025-12-11 16:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:20.552346952 +0000 UTC m=+98.232038418" watchObservedRunningTime="2025-12-11 16:54:20.552413624 +0000 UTC m=+98.232105090" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.598187 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.598227 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.598238 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.598251 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.598260 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.607073 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=14.607059373 podStartE2EDuration="14.607059373s" podCreationTimestamp="2025-12-11 16:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:20.606279064 +0000 UTC m=+98.285970530" watchObservedRunningTime="2025-12-11 16:54:20.607059373 +0000 UTC m=+98.286750839" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.700914 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.700975 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.700987 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.701004 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.701020 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.803506 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.803572 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.803593 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.803617 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.803638 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.901039 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:20 crc kubenswrapper[5109]: E1211 16:54:20.901284 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.901068 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:20 crc kubenswrapper[5109]: E1211 16:54:20.901999 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.903193 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:20 crc kubenswrapper[5109]: E1211 16:54:20.903343 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.903506 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:20 crc kubenswrapper[5109]: E1211 16:54:20.903684 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.906054 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.906093 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.906108 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.906127 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:20 crc kubenswrapper[5109]: I1211 16:54:20.906143 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:20Z","lastTransitionTime":"2025-12-11T16:54:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.008317 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.008376 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.008397 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.008417 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.008433 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.111625 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.111694 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.111715 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.111762 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.111781 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.214771 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.214865 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.214893 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.214928 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.214971 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.327296 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.327356 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.327376 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.327401 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.327419 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.356575 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" event={"ID":"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9","Type":"ContainerStarted","Data":"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.356657 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" event={"ID":"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9","Type":"ContainerStarted","Data":"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.363487 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.377691 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"3a14caf222afb62aaabdc47808b6f944","Type":"ContainerStarted","Data":"beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.379178 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.381355 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-rnsbf" event={"ID":"305d0e9b-83ad-424c-a5a5-c4ba12d3ae45","Type":"ContainerStarted","Data":"bbe8ac595d0665a1599d40294f6a8dc0b1d4c8ef3b5028ab9fcd4e9ec17a16ad"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.422655 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=15.422617871 podStartE2EDuration="15.422617871s" podCreationTimestamp="2025-12-11 16:54:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:21.413182963 +0000 UTC m=+99.092874469" watchObservedRunningTime="2025-12-11 16:54:21.422617871 +0000 UTC m=+99.102309367" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.438172 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-rnsbf" podStartSLOduration=79.438154786 podStartE2EDuration="1m19.438154786s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:21.437794997 +0000 UTC m=+99.117486463" watchObservedRunningTime="2025-12-11 16:54:21.438154786 +0000 UTC m=+99.117846262" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.439522 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.439568 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.439578 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.439591 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.439600 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.541223 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.541259 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.541268 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.541281 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.541290 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.643245 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.643284 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.643293 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.643306 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.643315 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.744848 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.744891 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.744900 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.744913 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.744922 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.755036 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.755093 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.755111 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.755135 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeNotReady" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.755164 5109 setters.go:618] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-11T16:54:21Z","lastTransitionTime":"2025-12-11T16:54:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.797652 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw"] Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.882733 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.884765 5109 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.886718 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\"" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.886921 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\"" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.890221 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"default-dockercfg-hqpm5\"" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.892688 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\"" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.896568 5109 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.983811 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0780697e-871e-4454-bc61-e09eee9567b0-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.985489 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0780697e-871e-4454-bc61-e09eee9567b0-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.985568 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0780697e-871e-4454-bc61-e09eee9567b0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.987538 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0780697e-871e-4454-bc61-e09eee9567b0-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:21 crc kubenswrapper[5109]: I1211 16:54:21.987642 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0780697e-871e-4454-bc61-e09eee9567b0-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.089124 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0780697e-871e-4454-bc61-e09eee9567b0-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.089181 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0780697e-871e-4454-bc61-e09eee9567b0-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.089437 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0780697e-871e-4454-bc61-e09eee9567b0-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.089502 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0780697e-871e-4454-bc61-e09eee9567b0-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.089528 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0780697e-871e-4454-bc61-e09eee9567b0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.089637 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/0780697e-871e-4454-bc61-e09eee9567b0-etc-cvo-updatepayloads\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.089644 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/0780697e-871e-4454-bc61-e09eee9567b0-etc-ssl-certs\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.090510 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0780697e-871e-4454-bc61-e09eee9567b0-service-ca\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.098211 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0780697e-871e-4454-bc61-e09eee9567b0-serving-cert\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.109649 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0780697e-871e-4454-bc61-e09eee9567b0-kube-api-access\") pod \"cluster-version-operator-7c9b9cfd6-blrgw\" (UID: \"0780697e-871e-4454-bc61-e09eee9567b0\") " pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.212975 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" Dec 11 16:54:22 crc kubenswrapper[5109]: W1211 16:54:22.225864 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0780697e_871e_4454_bc61_e09eee9567b0.slice/crio-1bdc8b561d2fa4182a67e97f26a4a40afd34262ef2d4c2bc8ad4b97fa6738d00 WatchSource:0}: Error finding container 1bdc8b561d2fa4182a67e97f26a4a40afd34262ef2d4c2bc8ad4b97fa6738d00: Status 404 returned error can't find the container with id 1bdc8b561d2fa4182a67e97f26a4a40afd34262ef2d4c2bc8ad4b97fa6738d00 Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.387210 5109 generic.go:358] "Generic (PLEG): container finished" podID="189d054c-6301-4cf4-ba60-da0380233d01" containerID="769e12a8f4ed4d75ed53fa9bf3eae701111821f02ea55c9edf5d6ff101321c4f" exitCode=0 Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.387331 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerDied","Data":"769e12a8f4ed4d75ed53fa9bf3eae701111821f02ea55c9edf5d6ff101321c4f"} Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.390442 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hx2kb" event={"ID":"7c45b4ae-02aa-4614-977b-544a9e303bdf","Type":"ContainerStarted","Data":"745ca6fee05059f7ae1cc1c5da3bbab828e24840a54cf61bb26296bf93ce4733"} Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.392689 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" event={"ID":"0780697e-871e-4454-bc61-e09eee9567b0","Type":"ContainerStarted","Data":"1bdc8b561d2fa4182a67e97f26a4a40afd34262ef2d4c2bc8ad4b97fa6738d00"} Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.594507 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.594574 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.594626 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.594787 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594805 5109 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594829 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594867 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594886 5109 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594921 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:38.594892559 +0000 UTC m=+116.274584055 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594973 5109 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594991 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.595037 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.595060 5109 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.594985 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:38.59495806 +0000 UTC m=+116.274649566 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.595175 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:38.595142435 +0000 UTC m=+116.274834001 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.595208 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:38.595191196 +0000 UTC m=+116.274882802 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.902514 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.902639 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.902957 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.903145 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.903613 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.903625 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.903733 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:22 crc kubenswrapper[5109]: E1211 16:54:22.903871 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.930141 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-hx2kb" podStartSLOduration=80.930093591 podStartE2EDuration="1m20.930093591s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:22.929013075 +0000 UTC m=+100.608704551" watchObservedRunningTime="2025-12-11 16:54:22.930093591 +0000 UTC m=+100.609785067" Dec 11 16:54:22 crc kubenswrapper[5109]: I1211 16:54:22.950090 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" podStartSLOduration=79.95006862299999 podStartE2EDuration="1m19.950068623s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:22.949443147 +0000 UTC m=+100.629134623" watchObservedRunningTime="2025-12-11 16:54:22.950068623 +0000 UTC m=+100.629760099" Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.002431 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.002613 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:23 crc kubenswrapper[5109]: E1211 16:54:23.003211 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:39.003189145 +0000 UTC m=+116.682880621 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:23 crc kubenswrapper[5109]: E1211 16:54:23.003328 5109 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:23 crc kubenswrapper[5109]: E1211 16:54:23.003398 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs podName:ad4ed271-a1a8-408d-b043-99cf044fb573 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:39.00337756 +0000 UTC m=+116.683069026 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs") pod "network-metrics-daemon-fp2cw" (UID: "ad4ed271-a1a8-408d-b043-99cf044fb573") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:23 crc kubenswrapper[5109]: E1211 16:54:23.086942 5109 cadvisor_stats_provider.go:525] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod561da642_b358_4ddc_86ee_669164d7b3c5.slice/crio-73bb9591f8cc88cb2850f8019795fc95ce1b6c4d1c220ccb45d0b565dafeb35a.scope\": RecentStats: unable to find data in memory cache]" Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.403702 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" event={"ID":"0780697e-871e-4454-bc61-e09eee9567b0","Type":"ContainerStarted","Data":"fe1cbc6c95b09862ca6a5af579b6d34f8bc598357e6f15049d093da21b351600"} Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.407555 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="73bb9591f8cc88cb2850f8019795fc95ce1b6c4d1c220ccb45d0b565dafeb35a" exitCode=0 Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.407629 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"73bb9591f8cc88cb2850f8019795fc95ce1b6c4d1c220ccb45d0b565dafeb35a"} Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.415831 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"11a2a3b9ed3d12072a480cd3571f89d3e3ee6c68f17bbd60c334530849f0b8ce"} Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.415895 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"f099a425a82983d885b200b73fcf89b1dba37bc85eba89c7f84a71efcc7178c7"} Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.418772 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerStarted","Data":"5f190d801dd01ceda80e2303c7ff346aed0409b71b377c77714a2f076072d606"} Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.420787 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-7bdcf4f5bd-7fjxv" event={"ID":"34177974-8d82-49d2-a763-391d0df3bbd8","Type":"ContainerStarted","Data":"20bdb5d81021559fdb5c0db09e70bd1e9eeb15e7b3b8e0e5b06547637748ca3c"} Dec 11 16:54:23 crc kubenswrapper[5109]: I1211 16:54:23.423413 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-7c9b9cfd6-blrgw" podStartSLOduration=81.423395889 podStartE2EDuration="1m21.423395889s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:23.422876636 +0000 UTC m=+101.102568132" watchObservedRunningTime="2025-12-11 16:54:23.423395889 +0000 UTC m=+101.103087385" Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.426717 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xqdz5" event={"ID":"61fc568b-d7c2-4633-b28c-f65428d01030","Type":"ContainerStarted","Data":"bd742dcdad191c1f3e88ed7a8b1d705c5c6c7482545459cf2a15b62b42bb0729"} Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.430313 5109 generic.go:358] "Generic (PLEG): container finished" podID="189d054c-6301-4cf4-ba60-da0380233d01" containerID="5f190d801dd01ceda80e2303c7ff346aed0409b71b377c77714a2f076072d606" exitCode=0 Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.430392 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerDied","Data":"5f190d801dd01ceda80e2303c7ff346aed0409b71b377c77714a2f076072d606"} Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.484232 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xqdz5" podStartSLOduration=81.484214117 podStartE2EDuration="1m21.484214117s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:24.446444895 +0000 UTC m=+102.126136381" watchObservedRunningTime="2025-12-11 16:54:24.484214117 +0000 UTC m=+102.163905583" Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.504635 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podStartSLOduration=82.504614989 podStartE2EDuration="1m22.504614989s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:24.501782471 +0000 UTC m=+102.181473987" watchObservedRunningTime="2025-12-11 16:54:24.504614989 +0000 UTC m=+102.184306465" Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.527104 5109 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.899280 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:24 crc kubenswrapper[5109]: E1211 16:54:24.899642 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.899328 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:24 crc kubenswrapper[5109]: E1211 16:54:24.899723 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.899342 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:24 crc kubenswrapper[5109]: E1211 16:54:24.899809 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:24 crc kubenswrapper[5109]: I1211 16:54:24.899287 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:24 crc kubenswrapper[5109]: E1211 16:54:24.899860 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:25 crc kubenswrapper[5109]: I1211 16:54:25.436603 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"ddb82e106ea206c6481e0e7996db55518f439ec5ea77c08b9b391f2eaced3ecf"} Dec 11 16:54:25 crc kubenswrapper[5109]: I1211 16:54:25.439169 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerStarted","Data":"e4a770a151ead9c9eb17ae89f70b4405782ec5ab448d4eea247e40e4d247b39b"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.444881 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-5jnd7" event={"ID":"428b39f5-eb1c-4f65-b7a4-eeb6e84860cc","Type":"ContainerStarted","Data":"634fdc3c094afcf1f105dd21ada3735104c5f56c2d636baa3d910733c3b109e5"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.451489 5109 generic.go:358] "Generic (PLEG): container finished" podID="189d054c-6301-4cf4-ba60-da0380233d01" containerID="e4a770a151ead9c9eb17ae89f70b4405782ec5ab448d4eea247e40e4d247b39b" exitCode=0 Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.451545 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerDied","Data":"e4a770a151ead9c9eb17ae89f70b4405782ec5ab448d4eea247e40e4d247b39b"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.455249 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"6f48c12cd8d9d7f9da38db4dd4293bdba9bd084e234c25324a74d281e52162eb"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.455303 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"c59ff5a852839e949601b2617bdae48e2624b8bef3e61979cbada802763747f6"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.455314 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"f56e7322033f70f60b9b525d77378a394971fe110d41e3f8ca772fc8ae86d4ed"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.455324 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"99d2321957866d3a56bdda8beca1e337cca93f57ad8fcdcae93e8787590c10a7"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.455333 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"068013b5c151d23b07bff0a7c4f87ba16ae9050d22993e0cb0474e1e32d7fe9b"} Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.899422 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.899456 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:26 crc kubenswrapper[5109]: E1211 16:54:26.899997 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.899575 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:26 crc kubenswrapper[5109]: I1211 16:54:26.899470 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:26 crc kubenswrapper[5109]: E1211 16:54:26.900153 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:26 crc kubenswrapper[5109]: E1211 16:54:26.900361 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:26 crc kubenswrapper[5109]: E1211 16:54:26.900537 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:28 crc kubenswrapper[5109]: I1211 16:54:28.463441 5109 generic.go:358] "Generic (PLEG): container finished" podID="189d054c-6301-4cf4-ba60-da0380233d01" containerID="78bb09c62b1127c3aae385975839e82a4a0171f30f8a808dba12cae66f404556" exitCode=0 Dec 11 16:54:28 crc kubenswrapper[5109]: I1211 16:54:28.463550 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerDied","Data":"78bb09c62b1127c3aae385975839e82a4a0171f30f8a808dba12cae66f404556"} Dec 11 16:54:28 crc kubenswrapper[5109]: I1211 16:54:28.902464 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:28 crc kubenswrapper[5109]: E1211 16:54:28.903106 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:28 crc kubenswrapper[5109]: I1211 16:54:28.902905 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:28 crc kubenswrapper[5109]: I1211 16:54:28.902950 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:28 crc kubenswrapper[5109]: I1211 16:54:28.902855 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:28 crc kubenswrapper[5109]: E1211 16:54:28.904261 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:28 crc kubenswrapper[5109]: E1211 16:54:28.904262 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:28 crc kubenswrapper[5109]: E1211 16:54:28.904037 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:29 crc kubenswrapper[5109]: I1211 16:54:29.471843 5109 generic.go:358] "Generic (PLEG): container finished" podID="189d054c-6301-4cf4-ba60-da0380233d01" containerID="1208142d3237f020c8d9fd405551f0117e42ef7b2eb9d42a9d1a50f38c07a1db" exitCode=0 Dec 11 16:54:29 crc kubenswrapper[5109]: I1211 16:54:29.472046 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerDied","Data":"1208142d3237f020c8d9fd405551f0117e42ef7b2eb9d42a9d1a50f38c07a1db"} Dec 11 16:54:29 crc kubenswrapper[5109]: I1211 16:54:29.478910 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"a9ecd830813d47e96969244bfd5812403f28e62d92f8d68d84eb4dae08215ad2"} Dec 11 16:54:30 crc kubenswrapper[5109]: I1211 16:54:30.900317 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:30 crc kubenswrapper[5109]: E1211 16:54:30.900566 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:30 crc kubenswrapper[5109]: I1211 16:54:30.900691 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:30 crc kubenswrapper[5109]: E1211 16:54:30.900894 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:30 crc kubenswrapper[5109]: I1211 16:54:30.901005 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:30 crc kubenswrapper[5109]: I1211 16:54:30.901005 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:30 crc kubenswrapper[5109]: E1211 16:54:30.901243 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:30 crc kubenswrapper[5109]: E1211 16:54:30.901438 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:32 crc kubenswrapper[5109]: I1211 16:54:32.899406 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:32 crc kubenswrapper[5109]: I1211 16:54:32.901998 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:32 crc kubenswrapper[5109]: I1211 16:54:32.902026 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:32 crc kubenswrapper[5109]: E1211 16:54:32.901981 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:32 crc kubenswrapper[5109]: I1211 16:54:32.902181 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:32 crc kubenswrapper[5109]: E1211 16:54:32.902361 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:32 crc kubenswrapper[5109]: E1211 16:54:32.902171 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:32 crc kubenswrapper[5109]: E1211 16:54:32.902533 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:32 crc kubenswrapper[5109]: I1211 16:54:32.916826 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:54:33 crc kubenswrapper[5109]: I1211 16:54:33.496052 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerStarted","Data":"b2feae0b9c2a623d8be88216569936659c684a4b49f64ec2b5aa9ce2c963a81d"} Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.505421 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerStarted","Data":"052d8728ba15ac15e54b397f9633ec810798f5f74e7189ec98767d0afdcf916a"} Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.506994 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.507063 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.507103 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.513828 5109 generic.go:358] "Generic (PLEG): container finished" podID="189d054c-6301-4cf4-ba60-da0380233d01" containerID="b2feae0b9c2a623d8be88216569936659c684a4b49f64ec2b5aa9ce2c963a81d" exitCode=0 Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.513915 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerDied","Data":"b2feae0b9c2a623d8be88216569936659c684a4b49f64ec2b5aa9ce2c963a81d"} Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.538951 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.549028 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.581798 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podStartSLOduration=92.581778839 podStartE2EDuration="1m32.581778839s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:34.554813639 +0000 UTC m=+112.234505115" watchObservedRunningTime="2025-12-11 16:54:34.581778839 +0000 UTC m=+112.261470305" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.899607 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.899629 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.899908 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:34 crc kubenswrapper[5109]: E1211 16:54:34.899912 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:34 crc kubenswrapper[5109]: E1211 16:54:34.900092 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:34 crc kubenswrapper[5109]: E1211 16:54:34.900228 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:34 crc kubenswrapper[5109]: I1211 16:54:34.900300 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:34 crc kubenswrapper[5109]: E1211 16:54:34.900403 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:36 crc kubenswrapper[5109]: I1211 16:54:36.899445 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:36 crc kubenswrapper[5109]: E1211 16:54:36.900285 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:36 crc kubenswrapper[5109]: I1211 16:54:36.899487 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:36 crc kubenswrapper[5109]: E1211 16:54:36.900450 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:36 crc kubenswrapper[5109]: I1211 16:54:36.899509 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:36 crc kubenswrapper[5109]: E1211 16:54:36.900588 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:36 crc kubenswrapper[5109]: I1211 16:54:36.899453 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:36 crc kubenswrapper[5109]: E1211 16:54:36.900776 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:37 crc kubenswrapper[5109]: I1211 16:54:37.531103 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" event={"ID":"189d054c-6301-4cf4-ba60-da0380233d01","Type":"ContainerStarted","Data":"7353182d0bc465dff007ddef2ea6ffdd5d5c09ac3bd70c197c5038d0e3bc968e"} Dec 11 16:54:37 crc kubenswrapper[5109]: I1211 16:54:37.567866 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-c2wzl" podStartSLOduration=95.567841843 podStartE2EDuration="1m35.567841843s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:37.566863339 +0000 UTC m=+115.246554865" watchObservedRunningTime="2025-12-11 16:54:37.567841843 +0000 UTC m=+115.247533349" Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.286989 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fp2cw"] Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.287147 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.287256 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.614367 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.614415 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.614438 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614553 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614568 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614578 5109 projected.go:194] Error preparing data for projected volume kube-api-access-l7w75 for pod openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614604 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614629 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75 podName:f863fff9-286a-45fa-b8f0-8a86994b8440 nodeName:}" failed. No retries permitted until 2025-12-11 16:55:10.614615372 +0000 UTC m=+148.294306838 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-l7w75" (UniqueName: "kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75") pod "network-check-source-5bb8f5cd97-xdvz5" (UID: "f863fff9-286a-45fa-b8f0-8a86994b8440") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614638 5109 projected.go:289] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614653 5109 projected.go:194] Error preparing data for projected volume kube-api-access-gwt8b for pod openshift-network-diagnostics/network-check-target-fhkjl: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614683 5109 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.614723 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b podName:17b87002-b798-480a-8e17-83053d698239 nodeName:}" failed. No retries permitted until 2025-12-11 16:55:10.614700344 +0000 UTC m=+148.294391810 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-gwt8b" (UniqueName: "kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b") pod "network-check-target-fhkjl" (UID: "17b87002-b798-480a-8e17-83053d698239") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.614948 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.615026 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:55:10.614993971 +0000 UTC m=+148.294685477 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.615046 5109 secret.go:189] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.615086 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert podName:6a9ae5f6-97bd-46ac-bafa-ca1b4452a141 nodeName:}" failed. No retries permitted until 2025-12-11 16:55:10.615075573 +0000 UTC m=+148.294767169 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert") pod "networking-console-plugin-5ff7774fd9-nljh6" (UID: "6a9ae5f6-97bd-46ac-bafa-ca1b4452a141") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.899413 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.899450 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.899713 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.899948 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:38 crc kubenswrapper[5109]: I1211 16:54:38.900045 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:38 crc kubenswrapper[5109]: E1211 16:54:38.900218 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:39 crc kubenswrapper[5109]: I1211 16:54:39.018555 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:39 crc kubenswrapper[5109]: E1211 16:54:39.018780 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:55:11.018713847 +0000 UTC m=+148.698405333 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:39 crc kubenswrapper[5109]: I1211 16:54:39.018889 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:39 crc kubenswrapper[5109]: E1211 16:54:39.019064 5109 secret.go:189] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:39 crc kubenswrapper[5109]: E1211 16:54:39.019180 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs podName:ad4ed271-a1a8-408d-b043-99cf044fb573 nodeName:}" failed. No retries permitted until 2025-12-11 16:55:11.019154258 +0000 UTC m=+148.698845734 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs") pod "network-metrics-daemon-fp2cw" (UID: "ad4ed271-a1a8-408d-b043-99cf044fb573") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 11 16:54:39 crc kubenswrapper[5109]: I1211 16:54:39.899708 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:39 crc kubenswrapper[5109]: E1211 16:54:39.899950 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:40 crc kubenswrapper[5109]: I1211 16:54:40.899669 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:40 crc kubenswrapper[5109]: I1211 16:54:40.899662 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:40 crc kubenswrapper[5109]: E1211 16:54:40.899825 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-fhkjl" podUID="17b87002-b798-480a-8e17-83053d698239" Dec 11 16:54:40 crc kubenswrapper[5109]: E1211 16:54:40.899927 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" podUID="6a9ae5f6-97bd-46ac-bafa-ca1b4452a141" Dec 11 16:54:40 crc kubenswrapper[5109]: I1211 16:54:40.899688 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:40 crc kubenswrapper[5109]: E1211 16:54:40.900103 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" podUID="f863fff9-286a-45fa-b8f0-8a86994b8440" Dec 11 16:54:41 crc kubenswrapper[5109]: I1211 16:54:41.899928 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:41 crc kubenswrapper[5109]: E1211 16:54:41.900131 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-fp2cw" podUID="ad4ed271-a1a8-408d-b043-99cf044fb573" Dec 11 16:54:42 crc kubenswrapper[5109]: I1211 16:54:42.658122 5109 kubelet_node_status.go:736] "Recording event message for node" node="crc" event="NodeReady" Dec 11 16:54:42 crc kubenswrapper[5109]: I1211 16:54:42.658377 5109 kubelet_node_status.go:550] "Fast updating node status as it just became ready" Dec 11 16:54:42 crc kubenswrapper[5109]: I1211 16:54:42.694053 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-9zkbs"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.441082 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.446632 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"client-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.446699 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.446821 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-djmfg\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.446909 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.446946 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.447694 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.448054 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-47r72"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.448160 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.448192 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.448602 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.451829 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.452486 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.452714 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.454678 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.455890 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.456401 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.456677 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-t8n29\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.458445 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-dgfs6"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.459188 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.465864 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.466623 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"etcd-serving-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.467222 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.467459 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"encryption-config-1\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.467537 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"audit-1\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.467655 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"etcd-client\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.468282 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.468421 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"image-import-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.468660 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.472250 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-4zqgh\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.473331 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-client-ca\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.473500 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-global-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.474958 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.475660 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-config\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.475714 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7efc5a6-365a-411a-8cc9-575086c1c169-serving-cert\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.475849 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.475882 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7efc5a6-365a-411a-8cc9-575086c1c169-tmp\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.475916 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psk95\" (UniqueName: \"kubernetes.io/projected/c7efc5a6-365a-411a-8cc9-575086c1c169-kube-api-access-psk95\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.477725 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.478329 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.480624 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/console-64d44f6ddf-hvlqh"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.480816 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.483351 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.484535 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.484584 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.484586 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.484995 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.485110 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-6n5ln\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.485255 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-images\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.488649 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.488980 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-tls\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.489855 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.490015 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.490220 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-rbac-proxy\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.491671 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.491844 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.491902 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.494042 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.494838 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.495132 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"audit-1\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.495169 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.495191 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-client\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.495473 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.496080 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.496406 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-mmcpt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.496418 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.496439 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-qqw4z\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.496457 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.496652 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.498893 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-52jsw"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.499267 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.499446 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.499828 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.500221 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"client-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.500412 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.505108 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-ws7p5"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.505339 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.506339 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"trusted-ca-bundle\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.507762 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.507803 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.507848 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.507949 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-747b44746d-bf7pr"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.508018 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.508062 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.508063 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.510932 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.511113 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-wzhvk\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.511231 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.511346 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.511462 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-8dkm8\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.511575 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.511962 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.512035 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.512139 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.512255 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-jmhxf\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.512360 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"config-operator-serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.512832 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.512946 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-sjn6s\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.513969 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"dns-operator-dockercfg-wbbsn\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.514067 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"metrics-tls\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.514149 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.515557 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-nkxtk"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.515605 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.515691 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-747b44746d-bf7pr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.516200 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.518000 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.518494 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-mdwwj\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.519701 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.519731 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-dockercfg-4vdnc\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.520077 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.520458 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.520702 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-ca-bundle\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.522116 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.522184 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.522246 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.523387 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-service-ca-bundle\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.528096 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-client\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.528468 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.528965 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6w67b\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.529296 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.530710 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.533182 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.536228 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"ingress-operator-dockercfg-74nwh\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.536447 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.536895 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"metrics-tls\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.537121 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.539277 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.539430 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.541162 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.547370 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-67c89758df-8rnw5"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.547874 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"trusted-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.548253 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.548864 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.549076 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-jcmfj\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.549251 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.550315 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.552973 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mr25z"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.553124 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.553481 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.555221 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.555316 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.557397 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.557504 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.559401 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.559542 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.562878 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.565498 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.565562 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.569140 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.569212 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.573763 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.573838 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576395 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smm9b\" (UniqueName: \"kubernetes.io/projected/2d87a27d-ab14-48d9-a685-c43beae0f998-kube-api-access-smm9b\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576480 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-encryption-config\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576540 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-psk95\" (UniqueName: \"kubernetes.io/projected/c7efc5a6-365a-411a-8cc9-575086c1c169-kube-api-access-psk95\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576573 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-oauth-config\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576618 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c6647e8-7cbf-4137-9bb6-2897bd47844c-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576728 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576820 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36cbe175-1eb2-4280-9656-fb631e70056c-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576844 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-image-import-ca\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576865 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-serving-cert\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.576922 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-etcd-client\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577035 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae8a431b-27bd-471a-b9d7-86211fc52f44-audit-dir\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577111 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74dcf\" (UniqueName: \"kubernetes.io/projected/ae8a431b-27bd-471a-b9d7-86211fc52f44-kube-api-access-74dcf\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577177 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-config\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577211 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkv55\" (UniqueName: \"kubernetes.io/projected/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-kube-api-access-jkv55\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577245 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0c6647e8-7cbf-4137-9bb6-2897bd47844c-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577277 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-trusted-ca-bundle\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577332 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-serving-cert\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577372 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-serving-cert\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577418 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/36cbe175-1eb2-4280-9656-fb631e70056c-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577442 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-etcd-client\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577460 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577536 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-service-ca\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577575 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7efc5a6-365a-411a-8cc9-575086c1c169-tmp\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577608 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psczc\" (UniqueName: \"kubernetes.io/projected/ce9e44ba-a2e2-443b-b265-aa312d55d7ca-kube-api-access-psczc\") pod \"downloads-747b44746d-bf7pr\" (UID: \"ce9e44ba-a2e2-443b-b265-aa312d55d7ca\") " pod="openshift-console/downloads-747b44746d-bf7pr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577633 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmh45\" (UniqueName: \"kubernetes.io/projected/357b32d1-3358-408f-915c-bc92802a28b9-kube-api-access-rmh45\") pod \"cluster-samples-operator-6b564684c8-84ffz\" (UID: \"357b32d1-3358-408f-915c-bc92802a28b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577659 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-config\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577678 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d87a27d-ab14-48d9-a685-c43beae0f998-audit-dir\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577717 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-audit-policies\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577753 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7efc5a6-365a-411a-8cc9-575086c1c169-serving-cert\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577796 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdm9t\" (UniqueName: \"kubernetes.io/projected/36cbe175-1eb2-4280-9656-fb631e70056c-kube-api-access-gdm9t\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577833 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-client-ca\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577857 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36cbe175-1eb2-4280-9656-fb631e70056c-config\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577876 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577911 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c398d16-ab6f-42b1-8f79-792bd8d2b637-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577933 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/357b32d1-3358-408f-915c-bc92802a28b9-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-84ffz\" (UID: \"357b32d1-3358-408f-915c-bc92802a28b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577961 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2d87a27d-ab14-48d9-a685-c43beae0f998-node-pullsecrets\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.577989 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-audit\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578011 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-config\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578030 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-oauth-serving-cert\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578054 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-trusted-ca-bundle\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578064 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7efc5a6-365a-411a-8cc9-575086c1c169-tmp\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578081 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0c6647e8-7cbf-4137-9bb6-2897bd47844c-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578102 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx2p9\" (UniqueName: \"kubernetes.io/projected/0c6647e8-7cbf-4137-9bb6-2897bd47844c-kube-api-access-dx2p9\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578123 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rlbq\" (UniqueName: \"kubernetes.io/projected/4c398d16-ab6f-42b1-8f79-792bd8d2b637-kube-api-access-2rlbq\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578147 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c398d16-ab6f-42b1-8f79-792bd8d2b637-images\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578166 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c398d16-ab6f-42b1-8f79-792bd8d2b637-config\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578185 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-etcd-serving-ca\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578208 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-encryption-config\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578514 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-proxy-ca-bundles\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578540 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-config\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578765 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578833 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-client-ca\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.578838 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.582658 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.582862 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.584713 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.585077 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7efc5a6-365a-411a-8cc9-575086c1c169-serving-cert\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.590454 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.590560 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.594514 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.594594 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.597613 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.597752 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.600401 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.600504 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.604082 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.604228 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.609422 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-tzg7t"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.609621 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.611246 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.613178 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-74545575db-8cq9r"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.613298 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.616324 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.616457 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.619898 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-68cf44c8b8-lth2t"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.620015 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.622756 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.622840 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.624523 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-kl6m8\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.625365 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-sqnwm"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.625560 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.628569 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-sjqx5"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.628652 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.632804 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.632876 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635683 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635710 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635722 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635732 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635757 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-9zkbs"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635769 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635778 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-ws7p5"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635789 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635799 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-47r72"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635808 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-dgfs6"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635818 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-747b44746d-bf7pr"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635828 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635839 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-nkxtk"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635849 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635860 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635869 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635880 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635889 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-67c89758df-8rnw5"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635899 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635908 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635917 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635850 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.635925 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636053 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-sjqx5"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636062 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636071 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d44f6ddf-hvlqh"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636079 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636088 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636097 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mr25z"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636106 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.636118 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-77chb"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.639775 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pp75t"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.640381 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643859 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-sqnwm"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643887 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-77chb"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643897 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643913 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643922 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-74545575db-8cq9r"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643932 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-52jsw"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643941 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643950 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pp75t"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643959 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-tzg7t"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643972 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643978 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.643982 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-nbqqf"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.644626 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.648542 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-4c2cr"] Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.648650 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.652907 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.664303 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.678998 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-oauth-config\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679048 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c6647e8-7cbf-4137-9bb6-2897bd47844c-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679266 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36cbe175-1eb2-4280-9656-fb631e70056c-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679297 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-image-import-ca\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679312 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-serving-cert\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679329 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-etcd-client\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679345 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae8a431b-27bd-471a-b9d7-86211fc52f44-audit-dir\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679361 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-74dcf\" (UniqueName: \"kubernetes.io/projected/ae8a431b-27bd-471a-b9d7-86211fc52f44-kube-api-access-74dcf\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679379 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jkv55\" (UniqueName: \"kubernetes.io/projected/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-kube-api-access-jkv55\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679398 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0c6647e8-7cbf-4137-9bb6-2897bd47844c-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679414 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-trusted-ca-bundle\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679430 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-serving-cert\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679446 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-serving-cert\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679467 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/36cbe175-1eb2-4280-9656-fb631e70056c-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679483 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-etcd-client\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679497 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679509 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ae8a431b-27bd-471a-b9d7-86211fc52f44-audit-dir\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679530 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-service-ca\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679562 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-psczc\" (UniqueName: \"kubernetes.io/projected/ce9e44ba-a2e2-443b-b265-aa312d55d7ca-kube-api-access-psczc\") pod \"downloads-747b44746d-bf7pr\" (UID: \"ce9e44ba-a2e2-443b-b265-aa312d55d7ca\") " pod="openshift-console/downloads-747b44746d-bf7pr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679588 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rmh45\" (UniqueName: \"kubernetes.io/projected/357b32d1-3358-408f-915c-bc92802a28b9-kube-api-access-rmh45\") pod \"cluster-samples-operator-6b564684c8-84ffz\" (UID: \"357b32d1-3358-408f-915c-bc92802a28b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679616 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-config\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679640 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d87a27d-ab14-48d9-a685-c43beae0f998-audit-dir\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679804 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d87a27d-ab14-48d9-a685-c43beae0f998-audit-dir\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679936 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-audit-policies\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.679991 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gdm9t\" (UniqueName: \"kubernetes.io/projected/36cbe175-1eb2-4280-9656-fb631e70056c-kube-api-access-gdm9t\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680057 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36cbe175-1eb2-4280-9656-fb631e70056c-config\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680092 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680122 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c398d16-ab6f-42b1-8f79-792bd8d2b637-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680143 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/357b32d1-3358-408f-915c-bc92802a28b9-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-84ffz\" (UID: \"357b32d1-3358-408f-915c-bc92802a28b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680170 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2d87a27d-ab14-48d9-a685-c43beae0f998-node-pullsecrets\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680198 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-audit\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680219 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-config\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680223 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-trusted-ca-bundle\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680243 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-oauth-serving-cert\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680294 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-trusted-ca-bundle\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680322 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0c6647e8-7cbf-4137-9bb6-2897bd47844c-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680340 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dx2p9\" (UniqueName: \"kubernetes.io/projected/0c6647e8-7cbf-4137-9bb6-2897bd47844c-kube-api-access-dx2p9\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680359 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2rlbq\" (UniqueName: \"kubernetes.io/projected/4c398d16-ab6f-42b1-8f79-792bd8d2b637-kube-api-access-2rlbq\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680384 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c398d16-ab6f-42b1-8f79-792bd8d2b637-images\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680403 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c398d16-ab6f-42b1-8f79-792bd8d2b637-config\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680419 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-etcd-serving-ca\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.680652 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-config\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681028 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0c6647e8-7cbf-4137-9bb6-2897bd47844c-trusted-ca\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681039 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-etcd-serving-ca\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681122 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-encryption-config\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681161 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-smm9b\" (UniqueName: \"kubernetes.io/projected/2d87a27d-ab14-48d9-a685-c43beae0f998-kube-api-access-smm9b\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681189 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-encryption-config\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681407 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ae8a431b-27bd-471a-b9d7-86211fc52f44-audit-policies\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681401 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-etcd-serving-ca\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681433 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/36cbe175-1eb2-4280-9656-fb631e70056c-tmp\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681515 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2d87a27d-ab14-48d9-a685-c43beae0f998-node-pullsecrets\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.681829 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4c398d16-ab6f-42b1-8f79-792bd8d2b637-config\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.682012 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-service-ca\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.682021 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-audit\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.682159 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-trusted-ca-bundle\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.682182 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4c398d16-ab6f-42b1-8f79-792bd8d2b637-images\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.682523 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-config\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.682945 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-image-import-ca\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.683248 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-etcd-client\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.683541 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36cbe175-1eb2-4280-9656-fb631e70056c-config\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.683721 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-oauth-serving-cert\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.684564 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d87a27d-ab14-48d9-a685-c43beae0f998-trusted-ca-bundle\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.685123 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-serving-cert\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.685626 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-encryption-config\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.686726 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-serving-cert\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.687093 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/357b32d1-3358-408f-915c-bc92802a28b9-samples-operator-tls\") pod \"cluster-samples-operator-6b564684c8-84ffz\" (UID: \"357b32d1-3358-408f-915c-bc92802a28b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.687543 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/ae8a431b-27bd-471a-b9d7-86211fc52f44-etcd-client\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.687616 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-serving-cert\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.689298 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d87a27d-ab14-48d9-a685-c43beae0f998-encryption-config\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.690188 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/4c398d16-ab6f-42b1-8f79-792bd8d2b637-machine-api-operator-tls\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.691298 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/0c6647e8-7cbf-4137-9bb6-2897bd47844c-metrics-tls\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.691410 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-console-oauth-config\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.695164 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/36cbe175-1eb2-4280-9656-fb631e70056c-serving-cert\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.705288 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"cluster-image-registry-operator-dockercfg-ntnd7\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.724150 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-operator-tls\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.763332 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-session\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.782390 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-client\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.782515 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52494b7-da4c-4aa1-82cc-285cc203d719-serving-cert\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.782591 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-client-ca\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.782681 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0562c782-fba6-47a1-87a5-69422d01de09-config\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.782729 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa291f60-d16f-476b-a0b0-0e988aa54e3b-metrics-tls\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.782856 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2sts\" (UniqueName: \"kubernetes.io/projected/aa291f60-d16f-476b-a0b0-0e988aa54e3b-kube-api-access-f2sts\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.782957 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0562c782-fba6-47a1-87a5-69422d01de09-machine-approver-tls\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.783013 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-certificates\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.783054 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eead6019-41c4-4ac6-9226-8ec4c456db13-installation-pull-secrets\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.783268 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e52494b7-da4c-4aa1-82cc-285cc203d719-tmp\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.783539 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-bound-sa-token\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.783690 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-serving-cert\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.783838 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.783932 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5zvn\" (UniqueName: \"kubernetes.io/projected/e52494b7-da4c-4aa1-82cc-285cc203d719-kube-api-access-v5zvn\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784032 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkkpm\" (UniqueName: \"kubernetes.io/projected/0562c782-fba6-47a1-87a5-69422d01de09-kube-api-access-pkkpm\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784072 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/637dcb89-3251-4807-8bf0-3c844ca1c05a-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784129 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-service-ca\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784193 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/637dcb89-3251-4807-8bf0-3c844ca1c05a-tmp\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784471 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/637dcb89-3251-4807-8bf0-3c844ca1c05a-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784501 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/637dcb89-3251-4807-8bf0-3c844ca1c05a-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784534 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/34a6df66-b508-4a7a-bc7e-254de76eaba7-tmp-dir\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784565 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlwt9\" (UniqueName: \"kubernetes.io/projected/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-kube-api-access-dlwt9\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784623 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8k4t\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-kube-api-access-r8k4t\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784712 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-config\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784776 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7zpl\" (UniqueName: \"kubernetes.io/projected/34a6df66-b508-4a7a-bc7e-254de76eaba7-kube-api-access-x7zpl\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784812 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-trusted-ca\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: E1211 16:54:43.784854 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.28482492 +0000 UTC m=+121.964516426 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.784428 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785028 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eead6019-41c4-4ac6-9226-8ec4c456db13-ca-trust-extracted\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785060 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34a6df66-b508-4a7a-bc7e-254de76eaba7-serving-cert\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785208 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xgl5\" (UniqueName: \"kubernetes.io/projected/637dcb89-3251-4807-8bf0-3c844ca1c05a-kube-api-access-6xgl5\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785371 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0562c782-fba6-47a1-87a5-69422d01de09-auth-proxy-config\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785518 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-ca\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785553 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-config\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785621 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-available-featuregates\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785648 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/637dcb89-3251-4807-8bf0-3c844ca1c05a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785680 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-tls\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.785769 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/aa291f60-d16f-476b-a0b0-0e988aa54e3b-tmp-dir\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.804535 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.841376 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-d2bf2\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.844628 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.883840 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"openshift-service-ca.crt\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.886358 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:43 crc kubenswrapper[5109]: E1211 16:54:43.886600 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.386556196 +0000 UTC m=+122.066247662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.886865 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f430caa3-dc3a-4ca1-8246-7606ef18a297-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.886911 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e52494b7-da4c-4aa1-82cc-285cc203d719-tmp\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.886958 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/099e943e-0a16-48e4-b19d-0cbf2dc39466-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.886985 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlkvm\" (UniqueName: \"kubernetes.io/projected/6b4042ae-cae4-4d16-a997-9c236defaa8e-kube-api-access-hlkvm\") pod \"package-server-manager-77f986bd66-7kbk8\" (UID: \"6b4042ae-cae4-4d16-a997-9c236defaa8e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887012 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/68e67f52-4e97-4950-8678-575d8fa94f99-profile-collector-cert\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887034 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btkf6\" (UniqueName: \"kubernetes.io/projected/17170d9a-42cd-4a85-bafb-b6ced9f3503c-kube-api-access-btkf6\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887060 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-bound-sa-token\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887146 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-csi-data-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887190 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-stats-auth\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887214 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v5zvn\" (UniqueName: \"kubernetes.io/projected/e52494b7-da4c-4aa1-82cc-285cc203d719-kube-api-access-v5zvn\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887262 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-certs\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887286 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxstg\" (UniqueName: \"kubernetes.io/projected/455d2dca-4769-4478-bbe5-ba72ed2db6ba-kube-api-access-nxstg\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887310 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-service-ca\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887335 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97vq8\" (UniqueName: \"kubernetes.io/projected/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-kube-api-access-97vq8\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887356 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-config\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887373 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887473 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887624 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887632 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e52494b7-da4c-4aa1-82cc-285cc203d719-tmp\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887685 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/637dcb89-3251-4807-8bf0-3c844ca1c05a-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887776 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6116425c-a3f6-4a09-91b1-207560b62779-serving-cert\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887829 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887865 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca35b32-bf7f-49d0-b796-060590d7c233-serving-cert\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887897 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/693595ff-e600-48ce-9d44-2f996cc27307-audit-dir\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887927 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.887969 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888014 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49tlz\" (UniqueName: \"kubernetes.io/projected/e707a0e8-be77-43fb-8236-19159201399a-kube-api-access-49tlz\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888070 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/584167c4-b6c8-4abb-80cd-7aaed553a463-apiservice-cert\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888107 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888157 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99387cea-c853-41ad-8926-39327379f170-secret-volume\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888193 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-config\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888220 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/68e67f52-4e97-4950-8678-575d8fa94f99-tmpfs\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888248 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-td7d8\" (UniqueName: \"kubernetes.io/projected/68e67f52-4e97-4950-8678-575d8fa94f99-kube-api-access-td7d8\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888273 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e707a0e8-be77-43fb-8236-19159201399a-images\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888312 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888341 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888383 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d56vm\" (UniqueName: \"kubernetes.io/projected/099e943e-0a16-48e4-b19d-0cbf2dc39466-kube-api-access-d56vm\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888500 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x7zpl\" (UniqueName: \"kubernetes.io/projected/34a6df66-b508-4a7a-bc7e-254de76eaba7-kube-api-access-x7zpl\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888556 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eca35b32-bf7f-49d0-b796-060590d7c233-kube-api-access\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888615 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/68e67f52-4e97-4950-8678-575d8fa94f99-srv-cert\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888713 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34a6df66-b508-4a7a-bc7e-254de76eaba7-serving-cert\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888782 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6xgl5\" (UniqueName: \"kubernetes.io/projected/637dcb89-3251-4807-8bf0-3c844ca1c05a-kube-api-access-6xgl5\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888821 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.888880 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f62vb\" (UniqueName: \"kubernetes.io/projected/99387cea-c853-41ad-8926-39327379f170-kube-api-access-f62vb\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889124 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/584167c4-b6c8-4abb-80cd-7aaed553a463-tmpfs\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889207 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c805a7f3-2074-4cfa-9001-aee570677c8e-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889266 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/28a5c354-8565-49aa-b329-cf529c594431-tmp\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889332 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8p7w\" (UniqueName: \"kubernetes.io/projected/73f65af5-9773-4165-8621-bf20d9eefd27-kube-api-access-g8p7w\") pod \"migrator-866fcbc849-n4k9x\" (UID: \"73f65af5-9773-4165-8621-bf20d9eefd27\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889407 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0562c782-fba6-47a1-87a5-69422d01de09-auth-proxy-config\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889437 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-ca\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889481 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-config\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889507 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q9qt\" (UniqueName: \"kubernetes.io/projected/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-kube-api-access-6q9qt\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890382 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh9rh\" (UniqueName: \"kubernetes.io/projected/2e862e11-c81e-4697-bee2-8cb2b0ba590e-kube-api-access-qh9rh\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890444 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/637dcb89-3251-4807-8bf0-3c844ca1c05a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890468 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6116425c-a3f6-4a09-91b1-207560b62779-config\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890491 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890531 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-tls\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890559 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4042ae-cae4-4d16-a997-9c236defaa8e-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-7kbk8\" (UID: \"6b4042ae-cae4-4d16-a997-9c236defaa8e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890584 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh8xp\" (UniqueName: \"kubernetes.io/projected/68d5a689-abae-41b3-8c8b-0a9255fc276c-kube-api-access-dh8xp\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890607 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8d2w\" (UniqueName: \"kubernetes.io/projected/1b24b31f-a9ae-4e65-88a9-6825ddff668e-kube-api-access-b8d2w\") pod \"multus-admission-controller-69db94689b-sqnwm\" (UID: \"1b24b31f-a9ae-4e65-88a9-6825ddff668e\") " pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890633 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c805a7f3-2074-4cfa-9001-aee570677c8e-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890657 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv59g\" (UniqueName: \"kubernetes.io/projected/28a5c354-8565-49aa-b329-cf529c594431-kube-api-access-zv59g\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890681 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-socket-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890712 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b125ace0-604e-48e1-808b-b7331df1c82a-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890790 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-client-ca\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890814 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f430caa3-dc3a-4ca1-8246-7606ef18a297-config\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890842 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e707a0e8-be77-43fb-8236-19159201399a-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890863 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c805a7f3-2074-4cfa-9001-aee570677c8e-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890888 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b125ace0-604e-48e1-808b-b7331df1c82a-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890933 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa291f60-d16f-476b-a0b0-0e988aa54e3b-metrics-tls\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890958 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f2sts\" (UniqueName: \"kubernetes.io/projected/aa291f60-d16f-476b-a0b0-0e988aa54e3b-kube-api-access-f2sts\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891008 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0562c782-fba6-47a1-87a5-69422d01de09-machine-approver-tls\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891030 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6116425c-a3f6-4a09-91b1-207560b62779-trusted-ca\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891059 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-certificates\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891083 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eead6019-41c4-4ac6-9226-8ec4c456db13-installation-pull-secrets\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891106 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/584167c4-b6c8-4abb-80cd-7aaed553a463-webhook-cert\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.889625 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-service-ca\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891134 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-8cmcd\" (UID: \"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890192 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0562c782-fba6-47a1-87a5-69422d01de09-auth-proxy-config\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891171 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-serving-cert\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891272 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq86z\" (UniqueName: \"kubernetes.io/projected/584167c4-b6c8-4abb-80cd-7aaed553a463-kube-api-access-nq86z\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891305 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891334 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p89gb\" (UniqueName: \"kubernetes.io/projected/5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7-kube-api-access-p89gb\") pod \"control-plane-machine-set-operator-75ffdb6fcd-8cmcd\" (UID: \"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891364 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pkkpm\" (UniqueName: \"kubernetes.io/projected/0562c782-fba6-47a1-87a5-69422d01de09-kube-api-access-pkkpm\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891367 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-config\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891389 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/637dcb89-3251-4807-8bf0-3c844ca1c05a-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.891565 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/34a6df66-b508-4a7a-bc7e-254de76eaba7-serving-cert\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.890331 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-ca\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892211 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892296 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-default-certificate\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892324 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/eca35b32-bf7f-49d0-b796-060590d7c233-tmp-dir\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892365 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892393 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/637dcb89-3251-4807-8bf0-3c844ca1c05a-tmp\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892422 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/637dcb89-3251-4807-8bf0-3c844ca1c05a-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892560 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-mountpoint-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892598 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/099e943e-0a16-48e4-b19d-0cbf2dc39466-tmpfs\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892661 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-node-bootstrap-token\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892683 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99387cea-c853-41ad-8926-39327379f170-config-volume\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892706 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-config\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892731 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/34a6df66-b508-4a7a-bc7e-254de76eaba7-tmp-dir\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892784 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca35b32-bf7f-49d0-b796-060590d7c233-config\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892905 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p58jx\" (UniqueName: \"kubernetes.io/projected/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-kube-api-access-p58jx\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892958 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dlwt9\" (UniqueName: \"kubernetes.io/projected/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-kube-api-access-dlwt9\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.892994 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-metrics-certs\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893031 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-r8k4t\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-kube-api-access-r8k4t\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893064 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/28a5c354-8565-49aa-b329-cf529c594431-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893099 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66s5r\" (UniqueName: \"kubernetes.io/projected/2494549f-2d12-462f-9afd-c11172476e1e-kube-api-access-66s5r\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893139 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mxgl\" (UniqueName: \"kubernetes.io/projected/b125ace0-604e-48e1-808b-b7331df1c82a-kube-api-access-4mxgl\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893171 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c805a7f3-2074-4cfa-9001-aee570677c8e-config\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893209 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-config\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893254 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455d2dca-4769-4478-bbe5-ba72ed2db6ba-config\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893311 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-trusted-ca\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893352 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/455d2dca-4769-4478-bbe5-ba72ed2db6ba-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893385 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-audit-policies\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893419 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/099e943e-0a16-48e4-b19d-0cbf2dc39466-srv-cert\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893461 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eead6019-41c4-4ac6-9226-8ec4c456db13-ca-trust-extracted\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893495 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893559 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c9jt5\" (UniqueName: \"kubernetes.io/projected/6116425c-a3f6-4a09-91b1-207560b62779-kube-api-access-c9jt5\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893599 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-ready\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893634 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1b24b31f-a9ae-4e65-88a9-6825ddff668e-webhook-certs\") pod \"multus-admission-controller-69db94689b-sqnwm\" (UID: \"1b24b31f-a9ae-4e65-88a9-6825ddff668e\") " pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.893676 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-cert\") pod \"ingress-canary-sjqx5\" (UID: \"d7d2f41c-6e62-4dba-b44a-982d0ce9e256\") " pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894072 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-available-featuregates\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894124 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-plugins-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894160 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894202 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/aa291f60-d16f-476b-a0b0-0e988aa54e3b-tmp-dir\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894241 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-key\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894273 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e862e11-c81e-4697-bee2-8cb2b0ba590e-service-ca-bundle\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894306 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17170d9a-42cd-4a85-bafb-b6ced9f3503c-metrics-tls\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894339 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/17170d9a-42cd-4a85-bafb-b6ced9f3503c-tmp-dir\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894379 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7d9j9\" (UniqueName: \"kubernetes.io/projected/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-kube-api-access-7d9j9\") pod \"ingress-canary-sjqx5\" (UID: \"d7d2f41c-6e62-4dba-b44a-982d0ce9e256\") " pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894426 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-client\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894462 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52494b7-da4c-4aa1-82cc-285cc203d719-serving-cert\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894506 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894542 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894575 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e707a0e8-be77-43fb-8236-19159201399a-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894618 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-serving-cert\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894655 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894701 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhdwz\" (UniqueName: \"kubernetes.io/projected/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-kube-api-access-mhdwz\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894764 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17170d9a-42cd-4a85-bafb-b6ced9f3503c-config-volume\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894807 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2494549f-2d12-462f-9afd-c11172476e1e-serving-cert\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894870 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-registration-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894910 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f430caa3-dc3a-4ca1-8246-7606ef18a297-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894945 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7xzzc\" (UniqueName: \"kubernetes.io/projected/693595ff-e600-48ce-9d44-2f996cc27307-kube-api-access-7xzzc\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.894988 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0562c782-fba6-47a1-87a5-69422d01de09-config\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.895021 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-cabundle\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.895054 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x86zc\" (UniqueName: \"kubernetes.io/projected/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-kube-api-access-x86zc\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.895092 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f430caa3-dc3a-4ca1-8246-7606ef18a297-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.895482 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/637dcb89-3251-4807-8bf0-3c844ca1c05a-trusted-ca\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.895698 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/637dcb89-3251-4807-8bf0-3c844ca1c05a-image-registry-operator-tls\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.895955 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/34a6df66-b508-4a7a-bc7e-254de76eaba7-tmp-dir\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.896114 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted-pem\" (UniqueName: \"kubernetes.io/empty-dir/637dcb89-3251-4807-8bf0-3c844ca1c05a-ca-trust-extracted-pem\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.896136 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/637dcb89-3251-4807-8bf0-3c844ca1c05a-tmp\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.896117 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/0562c782-fba6-47a1-87a5-69422d01de09-machine-approver-tls\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: E1211 16:54:43.896749 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.396722271 +0000 UTC m=+122.076413737 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.897023 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-available-featuregates\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.897382 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-tls\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.897470 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0562c782-fba6-47a1-87a5-69422d01de09-config\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.897673 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-client-ca\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.898065 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34a6df66-b508-4a7a-bc7e-254de76eaba7-config\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.898337 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/aa291f60-d16f-476b-a0b0-0e988aa54e3b-tmp-dir\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.900336 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eead6019-41c4-4ac6-9226-8ec4c456db13-ca-trust-extracted\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.901430 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-certificates\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.902344 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eead6019-41c4-4ac6-9226-8ec4c456db13-installation-pull-secrets\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.903084 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-serving-cert\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.903238 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-trusted-ca\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.904553 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.908385 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52494b7-da4c-4aa1-82cc-285cc203d719-serving-cert\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.911224 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/aa291f60-d16f-476b-a0b0-0e988aa54e3b-metrics-tls\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.913312 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/34a6df66-b508-4a7a-bc7e-254de76eaba7-etcd-client\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.923675 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.943356 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.963681 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.990703 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\"" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.995711 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:43 crc kubenswrapper[5109]: E1211 16:54:43.995846 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.495817644 +0000 UTC m=+122.175509150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.995933 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-csi-data-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.995979 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-stats-auth\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996020 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-certs\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996055 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-csi-data-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996054 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nxstg\" (UniqueName: \"kubernetes.io/projected/455d2dca-4769-4478-bbe5-ba72ed2db6ba-kube-api-access-nxstg\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996096 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-97vq8\" (UniqueName: \"kubernetes.io/projected/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-kube-api-access-97vq8\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996114 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-config\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996130 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996148 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996165 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6116425c-a3f6-4a09-91b1-207560b62779-serving-cert\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996185 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996200 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca35b32-bf7f-49d0-b796-060590d7c233-serving-cert\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996218 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/693595ff-e600-48ce-9d44-2f996cc27307-audit-dir\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996234 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996249 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996265 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-49tlz\" (UniqueName: \"kubernetes.io/projected/e707a0e8-be77-43fb-8236-19159201399a-kube-api-access-49tlz\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996281 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/584167c4-b6c8-4abb-80cd-7aaed553a463-apiservice-cert\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996296 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996312 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99387cea-c853-41ad-8926-39327379f170-secret-volume\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996327 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-config\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996347 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/68e67f52-4e97-4950-8678-575d8fa94f99-tmpfs\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996363 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-td7d8\" (UniqueName: \"kubernetes.io/projected/68e67f52-4e97-4950-8678-575d8fa94f99-kube-api-access-td7d8\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996383 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e707a0e8-be77-43fb-8236-19159201399a-images\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996399 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996415 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996431 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d56vm\" (UniqueName: \"kubernetes.io/projected/099e943e-0a16-48e4-b19d-0cbf2dc39466-kube-api-access-d56vm\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996448 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eca35b32-bf7f-49d0-b796-060590d7c233-kube-api-access\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996463 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/68e67f52-4e97-4950-8678-575d8fa94f99-srv-cert\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996487 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996504 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-f62vb\" (UniqueName: \"kubernetes.io/projected/99387cea-c853-41ad-8926-39327379f170-kube-api-access-f62vb\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996529 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/584167c4-b6c8-4abb-80cd-7aaed553a463-tmpfs\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996545 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c805a7f3-2074-4cfa-9001-aee570677c8e-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996562 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/28a5c354-8565-49aa-b329-cf529c594431-tmp\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996579 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g8p7w\" (UniqueName: \"kubernetes.io/projected/73f65af5-9773-4165-8621-bf20d9eefd27-kube-api-access-g8p7w\") pod \"migrator-866fcbc849-n4k9x\" (UID: \"73f65af5-9773-4165-8621-bf20d9eefd27\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996598 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6q9qt\" (UniqueName: \"kubernetes.io/projected/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-kube-api-access-6q9qt\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996614 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qh9rh\" (UniqueName: \"kubernetes.io/projected/2e862e11-c81e-4697-bee2-8cb2b0ba590e-kube-api-access-qh9rh\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996635 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6116425c-a3f6-4a09-91b1-207560b62779-config\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996651 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996668 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4042ae-cae4-4d16-a997-9c236defaa8e-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-7kbk8\" (UID: \"6b4042ae-cae4-4d16-a997-9c236defaa8e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996685 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dh8xp\" (UniqueName: \"kubernetes.io/projected/68d5a689-abae-41b3-8c8b-0a9255fc276c-kube-api-access-dh8xp\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996702 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-b8d2w\" (UniqueName: \"kubernetes.io/projected/1b24b31f-a9ae-4e65-88a9-6825ddff668e-kube-api-access-b8d2w\") pod \"multus-admission-controller-69db94689b-sqnwm\" (UID: \"1b24b31f-a9ae-4e65-88a9-6825ddff668e\") " pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996718 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c805a7f3-2074-4cfa-9001-aee570677c8e-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996748 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zv59g\" (UniqueName: \"kubernetes.io/projected/28a5c354-8565-49aa-b329-cf529c594431-kube-api-access-zv59g\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996765 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-socket-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996787 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b125ace0-604e-48e1-808b-b7331df1c82a-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996819 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f430caa3-dc3a-4ca1-8246-7606ef18a297-config\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996862 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e707a0e8-be77-43fb-8236-19159201399a-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996879 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c805a7f3-2074-4cfa-9001-aee570677c8e-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996897 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b125ace0-604e-48e1-808b-b7331df1c82a-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996928 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6116425c-a3f6-4a09-91b1-207560b62779-trusted-ca\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996949 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/584167c4-b6c8-4abb-80cd-7aaed553a463-webhook-cert\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996966 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-8cmcd\" (UID: \"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.996993 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nq86z\" (UniqueName: \"kubernetes.io/projected/584167c4-b6c8-4abb-80cd-7aaed553a463-kube-api-access-nq86z\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997014 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997030 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p89gb\" (UniqueName: \"kubernetes.io/projected/5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7-kube-api-access-p89gb\") pod \"control-plane-machine-set-operator-75ffdb6fcd-8cmcd\" (UID: \"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997050 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997068 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-default-certificate\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997084 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/eca35b32-bf7f-49d0-b796-060590d7c233-tmp-dir\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997103 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997123 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-mountpoint-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997140 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/099e943e-0a16-48e4-b19d-0cbf2dc39466-tmpfs\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997155 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-node-bootstrap-token\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997170 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99387cea-c853-41ad-8926-39327379f170-config-volume\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997188 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-config\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997205 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca35b32-bf7f-49d0-b796-060590d7c233-config\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997222 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-p58jx\" (UniqueName: \"kubernetes.io/projected/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-kube-api-access-p58jx\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997241 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-metrics-certs\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997259 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/28a5c354-8565-49aa-b329-cf529c594431-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997279 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-66s5r\" (UniqueName: \"kubernetes.io/projected/2494549f-2d12-462f-9afd-c11172476e1e-kube-api-access-66s5r\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997298 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4mxgl\" (UniqueName: \"kubernetes.io/projected/b125ace0-604e-48e1-808b-b7331df1c82a-kube-api-access-4mxgl\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997315 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c805a7f3-2074-4cfa-9001-aee570677c8e-config\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997340 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455d2dca-4769-4478-bbe5-ba72ed2db6ba-config\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997358 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/455d2dca-4769-4478-bbe5-ba72ed2db6ba-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997375 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-audit-policies\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997391 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/099e943e-0a16-48e4-b19d-0cbf2dc39466-srv-cert\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997408 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997429 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-c9jt5\" (UniqueName: \"kubernetes.io/projected/6116425c-a3f6-4a09-91b1-207560b62779-kube-api-access-c9jt5\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997444 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-ready\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997460 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1b24b31f-a9ae-4e65-88a9-6825ddff668e-webhook-certs\") pod \"multus-admission-controller-69db94689b-sqnwm\" (UID: \"1b24b31f-a9ae-4e65-88a9-6825ddff668e\") " pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997481 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-cert\") pod \"ingress-canary-sjqx5\" (UID: \"d7d2f41c-6e62-4dba-b44a-982d0ce9e256\") " pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997484 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/693595ff-e600-48ce-9d44-2f996cc27307-audit-dir\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997502 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-plugins-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997520 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997541 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-key\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997560 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e862e11-c81e-4697-bee2-8cb2b0ba590e-service-ca-bundle\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997577 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17170d9a-42cd-4a85-bafb-b6ced9f3503c-metrics-tls\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997593 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/17170d9a-42cd-4a85-bafb-b6ced9f3503c-tmp-dir\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997613 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7d9j9\" (UniqueName: \"kubernetes.io/projected/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-kube-api-access-7d9j9\") pod \"ingress-canary-sjqx5\" (UID: \"d7d2f41c-6e62-4dba-b44a-982d0ce9e256\") " pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997645 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997662 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997678 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e707a0e8-be77-43fb-8236-19159201399a-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997695 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-serving-cert\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997712 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997730 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mhdwz\" (UniqueName: \"kubernetes.io/projected/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-kube-api-access-mhdwz\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997758 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17170d9a-42cd-4a85-bafb-b6ced9f3503c-config-volume\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997777 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2494549f-2d12-462f-9afd-c11172476e1e-serving-cert\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997799 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-registration-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997815 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f430caa3-dc3a-4ca1-8246-7606ef18a297-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997833 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-7xzzc\" (UniqueName: \"kubernetes.io/projected/693595ff-e600-48ce-9d44-2f996cc27307-kube-api-access-7xzzc\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997852 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-cabundle\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997868 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-x86zc\" (UniqueName: \"kubernetes.io/projected/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-kube-api-access-x86zc\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997887 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f430caa3-dc3a-4ca1-8246-7606ef18a297-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997905 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f430caa3-dc3a-4ca1-8246-7606ef18a297-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997941 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/099e943e-0a16-48e4-b19d-0cbf2dc39466-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997959 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hlkvm\" (UniqueName: \"kubernetes.io/projected/6b4042ae-cae4-4d16-a997-9c236defaa8e-kube-api-access-hlkvm\") pod \"package-server-manager-77f986bd66-7kbk8\" (UID: \"6b4042ae-cae4-4d16-a997-9c236defaa8e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997977 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/68e67f52-4e97-4950-8678-575d8fa94f99-profile-collector-cert\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.997995 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-btkf6\" (UniqueName: \"kubernetes.io/projected/17170d9a-42cd-4a85-bafb-b6ced9f3503c-kube-api-access-btkf6\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.998251 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-service-ca\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.998370 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-socket-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.999101 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b125ace0-604e-48e1-808b-b7331df1c82a-mcc-auth-proxy-config\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:43 crc kubenswrapper[5109]: I1211 16:54:43.999362 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-ready\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:43.999467 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-plugins-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:43.999552 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/68e67f52-4e97-4950-8678-575d8fa94f99-tmpfs\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:43.999976 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-cliconfig\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.000317 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/17170d9a-42cd-4a85-bafb-b6ced9f3503c-tmp-dir\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.000419 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/584167c4-b6c8-4abb-80cd-7aaed553a463-tmpfs\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.000520 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.000882 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6116425c-a3f6-4a09-91b1-207560b62779-trusted-ca\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.000982 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/c805a7f3-2074-4cfa-9001-aee570677c8e-tmp-dir\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.001426 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/28a5c354-8565-49aa-b329-cf529c594431-tmp\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.001850 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/eca35b32-bf7f-49d0-b796-060590d7c233-tmp-dir\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.001959 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f430caa3-dc3a-4ca1-8246-7606ef18a297-tmp\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.002106 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.002168 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-error\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.001974 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-mountpoint-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.002524 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/68d5a689-abae-41b3-8c8b-0a9255fc276c-registration-dir\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.002616 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.003031 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6116425c-a3f6-4a09-91b1-207560b62779-config\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.003075 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/099e943e-0a16-48e4-b19d-0cbf2dc39466-tmpfs\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.003085 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e707a0e8-be77-43fb-8236-19159201399a-auth-proxy-config\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.003168 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-session\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.003297 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-router-certs\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.003482 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.503463318 +0000 UTC m=+122.183154824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.003686 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.004929 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6116425c-a3f6-4a09-91b1-207560b62779-serving-cert\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.005263 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-login\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.006028 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-serving-cert\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.006315 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.023474 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.036329 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.044077 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"audit\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.053470 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-audit-policies\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.063660 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.084279 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-6tbpn\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.100032 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.100244 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.600206264 +0000 UTC m=+122.279897770 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.101006 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.101436 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.601412552 +0000 UTC m=+122.281104058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.104881 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.117642 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2494549f-2d12-462f-9afd-c11172476e1e-serving-cert\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.123167 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-config\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.127323 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-config\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.151834 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"trusted-ca-bundle\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.162611 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-trusted-ca-bundle\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.163799 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"service-ca-bundle\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.173594 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2494549f-2d12-462f-9afd-c11172476e1e-service-ca-bundle\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.183429 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.203381 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.203672 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.203693 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.70365099 +0000 UTC m=+122.383342516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.204405 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.205292 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.70527814 +0000 UTC m=+122.384969596 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.223592 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.244031 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.251683 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-config\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.264695 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-6c46w\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.284558 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.291981 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-serving-cert\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.303717 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"openshift-service-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.306284 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.306560 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.806526234 +0000 UTC m=+122.486217740 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.307050 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.307710 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.807682892 +0000 UTC m=+122.487374398 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.324889 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.344960 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"pprof-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.353229 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99387cea-c853-41ad-8926-39327379f170-secret-volume\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.358264 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/099e943e-0a16-48e4-b19d-0cbf2dc39466-profile-collector-cert\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.359495 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/68e67f52-4e97-4950-8678-575d8fa94f99-profile-collector-cert\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.363783 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.374069 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99387cea-c853-41ad-8926-39327379f170-config-volume\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.384463 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.404087 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-tls\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.409717 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.409922 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.90989853 +0000 UTC m=+122.589589996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.410398 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.410706 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:44.910678318 +0000 UTC m=+122.590369824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.417609 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-75ffdb6fcd-8cmcd\" (UID: \"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.425148 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-dockercfg-gnx66\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.444563 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.465125 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-kknhg\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.485120 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.511495 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.511721 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.011698506 +0000 UTC m=+122.691389982 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.512938 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.513453 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.013437269 +0000 UTC m=+122.693128745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.521510 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-psk95\" (UniqueName: \"kubernetes.io/projected/c7efc5a6-365a-411a-8cc9-575086c1c169-kube-api-access-psk95\") pod \"controller-manager-65b6cccf98-9zkbs\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.524215 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"catalog-operator-serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.534889 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/099e943e-0a16-48e4-b19d-0cbf2dc39466-srv-cert\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.544004 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serviceaccount-dockercfg-4gqzj\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.564110 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.584401 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-scheduler-operator-serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.597134 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f430caa3-dc3a-4ca1-8246-7606ef18a297-serving-cert\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.602684 5109 request.go:752] "Waited before sending request" delay="1.018304061s" reason="client-side throttling, not priority and fairness" verb="GET" URL="https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-scheduler-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-kube-scheduler-operator-config&limit=500&resourceVersion=0" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.604164 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-config\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.613575 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f430caa3-dc3a-4ca1-8246-7606ef18a297-config\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.614240 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.614413 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.114392076 +0000 UTC m=+122.794083552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.614836 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.615175 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.115161074 +0000 UTC m=+122.794852540 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.625103 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-dockercfg-2wbn2\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.644034 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.655705 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/68e67f52-4e97-4950-8678-575d8fa94f99-srv-cert\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.665211 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-dockercfg-sw6nc\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.678283 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.685249 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mco-proxy-tls\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.695866 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e707a0e8-be77-43fb-8236-19159201399a-proxy-tls\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.704400 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-images\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.713847 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/e707a0e8-be77-43fb-8236-19159201399a-images\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.715789 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.716044 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.216022899 +0000 UTC m=+122.895714365 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.716297 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.716683 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.216676615 +0000 UTC m=+122.896368081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.723497 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"package-server-manager-serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.737588 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/6b4042ae-cae4-4d16-a997-9c236defaa8e-package-server-manager-serving-cert\") pod \"package-server-manager-77f986bd66-7kbk8\" (UID: \"6b4042ae-cae4-4d16-a997-9c236defaa8e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.744683 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.764294 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-dockercfg-tnfx9\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.783885 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-config\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.792694 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c805a7f3-2074-4cfa-9001-aee570677c8e-config\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.804728 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.815693 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c805a7f3-2074-4cfa-9001-aee570677c8e-serving-cert\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.820466 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.822413 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.322376176 +0000 UTC m=+123.002067692 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.824654 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-bjqfd\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.846573 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.858869 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-serving-cert\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.865158 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.872981 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-config\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.884511 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.912510 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.925064 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"packageserver-service-cert\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.925834 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.926518 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.4264768 +0000 UTC m=+123.106168326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.935479 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/584167c4-b6c8-4abb-80cd-7aaed553a463-webhook-cert\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.936413 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/584167c4-b6c8-4abb-80cd-7aaed553a463-apiservice-cert\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.946230 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.946981 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-9zkbs"] Dec 11 16:54:44 crc kubenswrapper[5109]: W1211 16:54:44.958068 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7efc5a6_365a_411a_8cc9_575086c1c169.slice/crio-c05d6e7f66922abdd01b49890d6eb4bc292f5f78eb7058bbba82e7f642c42794 WatchSource:0}: Error finding container c05d6e7f66922abdd01b49890d6eb4bc292f5f78eb7058bbba82e7f642c42794: Status 404 returned error can't find the container with id c05d6e7f66922abdd01b49890d6eb4bc292f5f78eb7058bbba82e7f642c42794 Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.964486 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-metrics\"" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.975918 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/28a5c354-8565-49aa-b329-cf529c594431-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:44 crc kubenswrapper[5109]: I1211 16:54:44.984091 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-dockercfg-2cfkp\"" Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997102 5109 secret.go:189] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997178 5109 secret.go:189] Couldn't get secret openshift-ingress/router-stats-default: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997195 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-certs podName:39e2e633-4eb3-4422-b9ef-5e81be4bff5e nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.497173746 +0000 UTC m=+123.176865222 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-certs") pod "machine-config-server-4c2cr" (UID: "39e2e633-4eb3-4422-b9ef-5e81be4bff5e") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997337 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-stats-auth podName:2e862e11-c81e-4697-bee2-8cb2b0ba590e nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.497310139 +0000 UTC m=+123.177001625 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "stats-auth" (UniqueName: "kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-stats-auth") pod "router-default-68cf44c8b8-lth2t" (UID: "2e862e11-c81e-4697-bee2-8cb2b0ba590e") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997375 5109 configmap.go:193] Couldn't get configMap openshift-marketplace/marketplace-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997424 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca podName:28a5c354-8565-49aa-b329-cf529c594431 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.497414622 +0000 UTC m=+123.177106098 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "marketplace-trusted-ca" (UniqueName: "kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca") pod "marketplace-operator-547dbd544d-tzg7t" (UID: "28a5c354-8565-49aa-b329-cf529c594431") : failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997642 5109 secret.go:189] Couldn't get secret openshift-kube-apiserver-operator/kube-apiserver-operator-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:44 crc kubenswrapper[5109]: E1211 16:54:44.997697 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eca35b32-bf7f-49d0-b796-060590d7c233-serving-cert podName:eca35b32-bf7f-49d0-b796-060590d7c233 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.497684859 +0000 UTC m=+123.177376395 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/eca35b32-bf7f-49d0-b796-060590d7c233-serving-cert") pod "kube-apiserver-operator-575994946d-crdzk" (UID: "eca35b32-bf7f-49d0-b796-060590d7c233") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000464 5109 secret.go:189] Couldn't get secret openshift-dns/dns-default-metrics-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000511 5109 secret.go:189] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000491 5109 secret.go:189] Couldn't get secret openshift-ingress-canary/canary-serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000589 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/17170d9a-42cd-4a85-bafb-b6ced9f3503c-metrics-tls podName:17170d9a-42cd-4a85-bafb-b6ced9f3503c nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.500542768 +0000 UTC m=+123.180234244 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-tls" (UniqueName: "kubernetes.io/secret/17170d9a-42cd-4a85-bafb-b6ced9f3503c-metrics-tls") pod "dns-default-77chb" (UID: "17170d9a-42cd-4a85-bafb-b6ced9f3503c") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000623 5109 configmap.go:193] Couldn't get configMap openshift-ingress/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000624 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-cert podName:d7d2f41c-6e62-4dba-b44a-982d0ce9e256 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.500602239 +0000 UTC m=+123.180293745 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-cert") pod "ingress-canary-sjqx5" (UID: "d7d2f41c-6e62-4dba-b44a-982d0ce9e256") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000655 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1b24b31f-a9ae-4e65-88a9-6825ddff668e-webhook-certs podName:1b24b31f-a9ae-4e65-88a9-6825ddff668e nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.50064262 +0000 UTC m=+123.180334096 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/1b24b31f-a9ae-4e65-88a9-6825ddff668e-webhook-certs") pod "multus-admission-controller-69db94689b-sqnwm" (UID: "1b24b31f-a9ae-4e65-88a9-6825ddff668e") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000683 5109 secret.go:189] Couldn't get secret openshift-service-ca/signing-key: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000690 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/2e862e11-c81e-4697-bee2-8cb2b0ba590e-service-ca-bundle podName:2e862e11-c81e-4697-bee2-8cb2b0ba590e nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.500680381 +0000 UTC m=+123.180371857 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/2e862e11-c81e-4697-bee2-8cb2b0ba590e-service-ca-bundle") pod "router-default-68cf44c8b8-lth2t" (UID: "2e862e11-c81e-4697-bee2-8cb2b0ba590e") : failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.000824 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-key podName:c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.500796934 +0000 UTC m=+123.180488410 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-key" (UniqueName: "kubernetes.io/secret/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-key") pod "service-ca-74545575db-8cq9r" (UID: "c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.001966 5109 configmap.go:193] Couldn't get configMap openshift-multus/cni-sysctl-allowlist: failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002016 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist podName:4eed92a0-6f24-424d-9be1-1f7e336ce8a9 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502004153 +0000 UTC m=+123.181695629 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cni-sysctl-allowlist" (UniqueName: "kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist") pod "cni-sysctl-allowlist-ds-nbqqf" (UID: "4eed92a0-6f24-424d-9be1-1f7e336ce8a9") : failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002020 5109 secret.go:189] Couldn't get secret openshift-ingress/router-certs-default: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002036 5109 secret.go:189] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002063 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-node-bootstrap-token podName:39e2e633-4eb3-4422-b9ef-5e81be4bff5e nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502055884 +0000 UTC m=+123.181747360 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-node-bootstrap-token") pod "machine-config-server-4c2cr" (UID: "39e2e633-4eb3-4422-b9ef-5e81be4bff5e") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002079 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-default-certificate podName:2e862e11-c81e-4697-bee2-8cb2b0ba590e nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502072105 +0000 UTC m=+123.181763581 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-certificate" (UniqueName: "kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-default-certificate") pod "router-default-68cf44c8b8-lth2t" (UID: "2e862e11-c81e-4697-bee2-8cb2b0ba590e") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002088 5109 secret.go:189] Couldn't get secret openshift-kube-storage-version-migrator-operator/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002106 5109 secret.go:189] Couldn't get secret openshift-ingress/router-metrics-certs-default: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002126 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/455d2dca-4769-4478-bbe5-ba72ed2db6ba-serving-cert podName:455d2dca-4769-4478-bbe5-ba72ed2db6ba nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502115586 +0000 UTC m=+123.181807062 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/455d2dca-4769-4478-bbe5-ba72ed2db6ba-serving-cert") pod "kube-storage-version-migrator-operator-565b79b866-l8msx" (UID: "455d2dca-4769-4478-bbe5-ba72ed2db6ba") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002147 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-metrics-certs podName:2e862e11-c81e-4697-bee2-8cb2b0ba590e nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502138576 +0000 UTC m=+123.181830182 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-metrics-certs") pod "router-default-68cf44c8b8-lth2t" (UID: "2e862e11-c81e-4697-bee2-8cb2b0ba590e") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002145 5109 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002182 5109 secret.go:189] Couldn't get secret openshift-machine-config-operator/mcc-proxy-tls: failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002209 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b125ace0-604e-48e1-808b-b7331df1c82a-proxy-tls podName:b125ace0-604e-48e1-808b-b7331df1c82a nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502202778 +0000 UTC m=+123.181894254 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-tls" (UniqueName: "kubernetes.io/secret/b125ace0-604e-48e1-808b-b7331df1c82a-proxy-tls") pod "machine-config-controller-f9cdd68f7-xzb7d" (UID: "b125ace0-604e-48e1-808b-b7331df1c82a") : failed to sync secret cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002238 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/eca35b32-bf7f-49d0-b796-060590d7c233-config podName:eca35b32-bf7f-49d0-b796-060590d7c233 nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502220118 +0000 UTC m=+123.181911594 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/eca35b32-bf7f-49d0-b796-060590d7c233-config") pod "kube-apiserver-operator-575994946d-crdzk" (UID: "eca35b32-bf7f-49d0-b796-060590d7c233") : failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002239 5109 configmap.go:193] Couldn't get configMap openshift-kube-storage-version-migrator-operator/config: failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002257 5109 configmap.go:193] Couldn't get configMap openshift-service-ca/signing-cabundle: failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002277 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/455d2dca-4769-4478-bbe5-ba72ed2db6ba-config podName:455d2dca-4769-4478-bbe5-ba72ed2db6ba nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502268619 +0000 UTC m=+123.181960115 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/455d2dca-4769-4478-bbe5-ba72ed2db6ba-config") pod "kube-storage-version-migrator-operator-565b79b866-l8msx" (UID: "455d2dca-4769-4478-bbe5-ba72ed2db6ba") : failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002291 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-cabundle podName:c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.50228418 +0000 UTC m=+123.181975646 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "signing-cabundle" (UniqueName: "kubernetes.io/configmap/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-cabundle") pod "service-ca-74545575db-8cq9r" (UID: "c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd") : failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002842 5109 configmap.go:193] Couldn't get configMap openshift-dns/dns-default: failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.002932 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/17170d9a-42cd-4a85-bafb-b6ced9f3503c-config-volume podName:17170d9a-42cd-4a85-bafb-b6ced9f3503c nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.502905194 +0000 UTC m=+123.182596700 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/17170d9a-42cd-4a85-bafb-b6ced9f3503c-config-volume") pod "dns-default-77chb" (UID: "17170d9a-42cd-4a85-bafb-b6ced9f3503c") : failed to sync configmap cache: timed out waiting for the condition Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.012303 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"marketplace-trusted-ca\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.023515 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.028735 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.028858 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.528836111 +0000 UTC m=+123.208527587 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.029201 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.029868 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.529831514 +0000 UTC m=+123.209523020 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.044113 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.064041 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-bgxvm\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.085085 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.104648 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.124174 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.131274 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.131522 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.631443257 +0000 UTC m=+123.311134733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.132502 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.132871 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.632858132 +0000 UTC m=+123.312549598 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.144402 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mcc-proxy-tls\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.164842 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-controller-dockercfg-xnj77\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.185451 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-certs-default\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.204796 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.224733 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.235760 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.236206 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.736178286 +0000 UTC m=+123.415869822 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.237672 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.238305 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.738265476 +0000 UTC m=+123.417956972 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.245239 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.264853 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.284681 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-kw8fx\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.305206 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.324238 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-2h6bs\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.339775 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.339907 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.839876579 +0000 UTC m=+123.519568055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.340373 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.340668 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.840655167 +0000 UTC m=+123.520346633 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.344003 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.379199 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.383719 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.403835 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.425383 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ac-dockercfg-gj7jx\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.441410 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.441548 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.941518233 +0000 UTC m=+123.621209739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.442374 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.442994 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:45.942966868 +0000 UTC m=+123.622658374 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.445153 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-admission-controller-secret\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.465109 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9pgs7\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.484427 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.504405 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.524869 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.543802 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.544116 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.044077079 +0000 UTC m=+123.723768585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.544557 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-node-bootstrap-token\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.544638 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca35b32-bf7f-49d0-b796-060590d7c233-config\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.544898 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-metrics-certs\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545036 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455d2dca-4769-4478-bbe5-ba72ed2db6ba-config\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545099 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/455d2dca-4769-4478-bbe5-ba72ed2db6ba-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545162 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1b24b31f-a9ae-4e65-88a9-6825ddff668e-webhook-certs\") pod \"multus-admission-controller-69db94689b-sqnwm\" (UID: \"1b24b31f-a9ae-4e65-88a9-6825ddff668e\") " pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545207 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-cert\") pod \"ingress-canary-sjqx5\" (UID: \"d7d2f41c-6e62-4dba-b44a-982d0ce9e256\") " pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545848 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-key\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545883 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e862e11-c81e-4697-bee2-8cb2b0ba590e-service-ca-bundle\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545904 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17170d9a-42cd-4a85-bafb-b6ced9f3503c-metrics-tls\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545941 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17170d9a-42cd-4a85-bafb-b6ced9f3503c-config-volume\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.545971 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-cabundle\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546025 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-stats-auth\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546055 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-certs\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546094 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546119 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca35b32-bf7f-49d0-b796-060590d7c233-serving-cert\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546257 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b125ace0-604e-48e1-808b-b7331df1c82a-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546365 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546407 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-default-certificate\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546438 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.546995 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/455d2dca-4769-4478-bbe5-ba72ed2db6ba-config\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.547599 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-dockercfg-bf7fj\"" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.548587 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.048549516 +0000 UTC m=+123.728241032 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.549587 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-cabundle\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.549680 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2e862e11-c81e-4697-bee2-8cb2b0ba590e-service-ca-bundle\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.550092 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/1b24b31f-a9ae-4e65-88a9-6825ddff668e-webhook-certs\") pod \"multus-admission-controller-69db94689b-sqnwm\" (UID: \"1b24b31f-a9ae-4e65-88a9-6825ddff668e\") " pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.551900 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-signing-key\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.552182 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.552339 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-metrics-certs\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.552456 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b125ace0-604e-48e1-808b-b7331df1c82a-proxy-tls\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.554039 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-cert\") pod \"ingress-canary-sjqx5\" (UID: \"d7d2f41c-6e62-4dba-b44a-982d0ce9e256\") " pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.555529 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-stats-auth\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.556607 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/455d2dca-4769-4478-bbe5-ba72ed2db6ba-serving-cert\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.560367 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" event={"ID":"c7efc5a6-365a-411a-8cc9-575086c1c169","Type":"ContainerStarted","Data":"4b6fbed6af02d089962b5aebe82e539ad168a4c742fc2ff980cdbf0e3be91aef"} Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.560428 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" event={"ID":"c7efc5a6-365a-411a-8cc9-575086c1c169","Type":"ContainerStarted","Data":"c05d6e7f66922abdd01b49890d6eb4bc292f5f78eb7058bbba82e7f642c42794"} Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.561583 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.561853 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/2e862e11-c81e-4697-bee2-8cb2b0ba590e-default-certificate\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.564552 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-root-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.585528 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-serving-cert\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.596381 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eca35b32-bf7f-49d0-b796-060590d7c233-serving-cert\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.604652 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-config\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.605961 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eca35b32-bf7f-49d0-b796-060590d7c233-config\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.622920 5109 request.go:752] "Waited before sending request" delay="1.982263422s" reason="client-side throttling, not priority and fairness" verb="GET" URL="https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/configmaps?fieldSelector=metadata.name%3Ddns-default&limit=500&resourceVersion=0" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.627574 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.629582 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/17170d9a-42cd-4a85-bafb-b6ced9f3503c-config-volume\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.645473 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-kpvmz\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.648443 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.648902 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.148865458 +0000 UTC m=+123.828556964 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.664910 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.674056 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/17170d9a-42cd-4a85-bafb-b6ced9f3503c-metrics-tls\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.684426 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"kube-root-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.704495 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"openshift-service-ca.crt\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.724388 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"hostpath-provisioner\"/\"csi-hostpath-provisioner-sa-dockercfg-7dcws\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.744174 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-sysctl-allowlist\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.748720 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.752517 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.752872 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.252856208 +0000 UTC m=+123.932547754 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.764554 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"node-bootstrapper-token\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.770082 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-node-bootstrap-token\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.784494 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-tls\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.792170 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-certs\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.805427 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-dockercfg-dzw6b\"" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.850439 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmh45\" (UniqueName: \"kubernetes.io/projected/357b32d1-3358-408f-915c-bc92802a28b9-kube-api-access-rmh45\") pod \"cluster-samples-operator-6b564684c8-84ffz\" (UID: \"357b32d1-3358-408f-915c-bc92802a28b9\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.854075 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.854353 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.354316657 +0000 UTC m=+124.034008133 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.854580 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.855528 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.355515156 +0000 UTC m=+124.035206622 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.858450 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-74dcf\" (UniqueName: \"kubernetes.io/projected/ae8a431b-27bd-471a-b9d7-86211fc52f44-kube-api-access-74dcf\") pod \"apiserver-8596bd845d-gb4zv\" (UID: \"ae8a431b-27bd-471a-b9d7-86211fc52f44\") " pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.877436 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.879602 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-psczc\" (UniqueName: \"kubernetes.io/projected/ce9e44ba-a2e2-443b-b265-aa312d55d7ca-kube-api-access-psczc\") pod \"downloads-747b44746d-bf7pr\" (UID: \"ce9e44ba-a2e2-443b-b265-aa312d55d7ca\") " pod="openshift-console/downloads-747b44746d-bf7pr" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.899448 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx2p9\" (UniqueName: \"kubernetes.io/projected/0c6647e8-7cbf-4137-9bb6-2897bd47844c-kube-api-access-dx2p9\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.916683 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkv55\" (UniqueName: \"kubernetes.io/projected/e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d-kube-api-access-jkv55\") pod \"console-64d44f6ddf-hvlqh\" (UID: \"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d\") " pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.937673 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rlbq\" (UniqueName: \"kubernetes.io/projected/4c398d16-ab6f-42b1-8f79-792bd8d2b637-kube-api-access-2rlbq\") pod \"machine-api-operator-755bb95488-dgfs6\" (UID: \"4c398d16-ab6f-42b1-8f79-792bd8d2b637\") " pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.949805 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.955599 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.955825 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.455797267 +0000 UTC m=+124.135488743 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.956167 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:45 crc kubenswrapper[5109]: E1211 16:54:45.956556 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.456544265 +0000 UTC m=+124.136235741 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.957532 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.963778 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdm9t\" (UniqueName: \"kubernetes.io/projected/36cbe175-1eb2-4280-9656-fb631e70056c-kube-api-access-gdm9t\") pod \"openshift-controller-manager-operator-686468bdd5-xr7js\" (UID: \"36cbe175-1eb2-4280-9656-fb631e70056c\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.971975 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.978534 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-smm9b\" (UniqueName: \"kubernetes.io/projected/2d87a27d-ab14-48d9-a685-c43beae0f998-kube-api-access-smm9b\") pod \"apiserver-9ddfb9f55-47r72\" (UID: \"2d87a27d-ab14-48d9-a685-c43beae0f998\") " pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:45 crc kubenswrapper[5109]: I1211 16:54:45.985359 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.006101 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0c6647e8-7cbf-4137-9bb6-2897bd47844c-bound-sa-token\") pod \"ingress-operator-6b9cb4dbcf-zx6sp\" (UID: \"0c6647e8-7cbf-4137-9bb6-2897bd47844c\") " pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.012876 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-747b44746d-bf7pr" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.036884 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.039521 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-bound-sa-token\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.044997 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.057560 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.057772 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.557731978 +0000 UTC m=+124.237423444 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.058027 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.058411 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.558393634 +0000 UTC m=+124.238085100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.073565 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/637dcb89-3251-4807-8bf0-3c844ca1c05a-bound-sa-token\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.079053 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5zvn\" (UniqueName: \"kubernetes.io/projected/e52494b7-da4c-4aa1-82cc-285cc203d719-kube-api-access-v5zvn\") pod \"route-controller-manager-776cdc94d6-gp4qk\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.105683 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7zpl\" (UniqueName: \"kubernetes.io/projected/34a6df66-b508-4a7a-bc7e-254de76eaba7-kube-api-access-x7zpl\") pod \"etcd-operator-69b85846b6-6jvnh\" (UID: \"34a6df66-b508-4a7a-bc7e-254de76eaba7\") " pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.126544 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xgl5\" (UniqueName: \"kubernetes.io/projected/637dcb89-3251-4807-8bf0-3c844ca1c05a-kube-api-access-6xgl5\") pod \"cluster-image-registry-operator-86c45576b9-njh8k\" (UID: \"637dcb89-3251-4807-8bf0-3c844ca1c05a\") " pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.142479 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2sts\" (UniqueName: \"kubernetes.io/projected/aa291f60-d16f-476b-a0b0-0e988aa54e3b-kube-api-access-f2sts\") pod \"dns-operator-799b87ffcd-ws7p5\" (UID: \"aa291f60-d16f-476b-a0b0-0e988aa54e3b\") " pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.160100 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8k4t\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-kube-api-access-r8k4t\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.162136 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.162565 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.662552747 +0000 UTC m=+124.342244213 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.183662 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkkpm\" (UniqueName: \"kubernetes.io/projected/0562c782-fba6-47a1-87a5-69422d01de09-kube-api-access-pkkpm\") pod \"machine-approver-54c688565-5lz5m\" (UID: \"0562c782-fba6-47a1-87a5-69422d01de09\") " pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.197488 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlwt9\" (UniqueName: \"kubernetes.io/projected/4cb7d92e-f9a5-4deb-a375-7f48da3f0899-kube-api-access-dlwt9\") pod \"openshift-config-operator-5777786469-52jsw\" (UID: \"4cb7d92e-f9a5-4deb-a375-7f48da3f0899\") " pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.217454 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxstg\" (UniqueName: \"kubernetes.io/projected/455d2dca-4769-4478-bbe5-ba72ed2db6ba-kube-api-access-nxstg\") pod \"kube-storage-version-migrator-operator-565b79b866-l8msx\" (UID: \"455d2dca-4769-4478-bbe5-ba72ed2db6ba\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.229128 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.238646 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.239155 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-97vq8\" (UniqueName: \"kubernetes.io/projected/39e2e633-4eb3-4422-b9ef-5e81be4bff5e-kube-api-access-97vq8\") pod \"machine-config-server-4c2cr\" (UID: \"39e2e633-4eb3-4422-b9ef-5e81be4bff5e\") " pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.258286 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.258348 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-747b44746d-bf7pr"] Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.260926 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c6647e8_7cbf_4137_9bb6_2897bd47844c.slice/crio-512770fc5c57d85549f8300bf06ead95aa7b30135d0a05795769e8610ea3dc3b WatchSource:0}: Error finding container 512770fc5c57d85549f8300bf06ead95aa7b30135d0a05795769e8610ea3dc3b: Status 404 returned error can't find the container with id 512770fc5c57d85549f8300bf06ead95aa7b30135d0a05795769e8610ea3dc3b Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.261916 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podce9e44ba_a2e2_443b_b265_aa312d55d7ca.slice/crio-71a78348787c9cc57edf314dee82a6e89ea1545164870f0c9714df6d9cf40534 WatchSource:0}: Error finding container 71a78348787c9cc57edf314dee82a6e89ea1545164870f0c9714df6d9cf40534: Status 404 returned error can't find the container with id 71a78348787c9cc57edf314dee82a6e89ea1545164870f0c9714df6d9cf40534 Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.265112 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.265262 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8d2w\" (UniqueName: \"kubernetes.io/projected/1b24b31f-a9ae-4e65-88a9-6825ddff668e-kube-api-access-b8d2w\") pod \"multus-admission-controller-69db94689b-sqnwm\" (UID: \"1b24b31f-a9ae-4e65-88a9-6825ddff668e\") " pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.267058 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.76704561 +0000 UTC m=+124.446737076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.267349 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.278168 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.287886 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.288176 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-btkf6\" (UniqueName: \"kubernetes.io/projected/17170d9a-42cd-4a85-bafb-b6ced9f3503c-kube-api-access-btkf6\") pod \"dns-default-77chb\" (UID: \"17170d9a-42cd-4a85-bafb-b6ced9f3503c\") " pod="openshift-dns/dns-default-77chb" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.299090 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.299126 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c805a7f3-2074-4cfa-9001-aee570677c8e-kube-api-access\") pod \"kube-controller-manager-operator-69d5f845f8-lcqcm\" (UID: \"c805a7f3-2074-4cfa-9001-aee570677c8e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.305717 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.312246 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36cbe175_1eb2_4280_9656_fb631e70056c.slice/crio-a5a71b9f44d2adb31493d52a5981af98c983c932e6e07227c6b02ae1ffe4505a WatchSource:0}: Error finding container a5a71b9f44d2adb31493d52a5981af98c983c932e6e07227c6b02ae1ffe4505a: Status 404 returned error can't find the container with id a5a71b9f44d2adb31493d52a5981af98c983c932e6e07227c6b02ae1ffe4505a Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.317910 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.319917 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.321831 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv59g\" (UniqueName: \"kubernetes.io/projected/28a5c354-8565-49aa-b329-cf529c594431-kube-api-access-zv59g\") pod \"marketplace-operator-547dbd544d-tzg7t\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.327222 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.342291 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-49tlz\" (UniqueName: \"kubernetes.io/projected/e707a0e8-be77-43fb-8236-19159201399a-kube-api-access-49tlz\") pod \"machine-config-operator-67c9d58cbb-9hbcp\" (UID: \"e707a0e8-be77-43fb-8236-19159201399a\") " pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.357032 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.357879 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-77chb" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.364069 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-c9jt5\" (UniqueName: \"kubernetes.io/projected/6116425c-a3f6-4a09-91b1-207560b62779-kube-api-access-c9jt5\") pod \"console-operator-67c89758df-8rnw5\" (UID: \"6116425c-a3f6-4a09-91b1-207560b62779\") " pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.368416 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.369055 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.869031992 +0000 UTC m=+124.548723468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.382928 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-64d44f6ddf-hvlqh"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.384286 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.391708 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/eca35b32-bf7f-49d0-b796-060590d7c233-kube-api-access\") pod \"kube-apiserver-operator-575994946d-crdzk\" (UID: \"eca35b32-bf7f-49d0-b796-060590d7c233\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.393477 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-755bb95488-dgfs6"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.400563 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-4c2cr" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.407006 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh9rh\" (UniqueName: \"kubernetes.io/projected/2e862e11-c81e-4697-bee2-8cb2b0ba590e-kube-api-access-qh9rh\") pod \"router-default-68cf44c8b8-lth2t\" (UID: \"2e862e11-c81e-4697-bee2-8cb2b0ba590e\") " pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.412385 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4dc2bdf_6e49_4012_9f1c_d2af161dfd6d.slice/crio-cdab104672ec63ae7453f0b5f21868f40d0b02ee23af47f64cd1d42c9be517da WatchSource:0}: Error finding container cdab104672ec63ae7453f0b5f21868f40d0b02ee23af47f64cd1d42c9be517da: Status 404 returned error can't find the container with id cdab104672ec63ae7453f0b5f21868f40d0b02ee23af47f64cd1d42c9be517da Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.445523 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7d9j9\" (UniqueName: \"kubernetes.io/projected/d7d2f41c-6e62-4dba-b44a-982d0ce9e256-kube-api-access-7d9j9\") pod \"ingress-canary-sjqx5\" (UID: \"d7d2f41c-6e62-4dba-b44a-982d0ce9e256\") " pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.451471 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d56vm\" (UniqueName: \"kubernetes.io/projected/099e943e-0a16-48e4-b19d-0cbf2dc39466-kube-api-access-d56vm\") pod \"catalog-operator-75ff9f647d-ptnqv\" (UID: \"099e943e-0a16-48e4-b19d-0cbf2dc39466\") " pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.454498 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-9ddfb9f55-47r72"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.467167 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.467512 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8p7w\" (UniqueName: \"kubernetes.io/projected/73f65af5-9773-4165-8621-bf20d9eefd27-kube-api-access-g8p7w\") pod \"migrator-866fcbc849-n4k9x\" (UID: \"73f65af5-9773-4165-8621-bf20d9eefd27\") " pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.470450 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.470875 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:46.97086215 +0000 UTC m=+124.650553616 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.488394 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-f62vb\" (UniqueName: \"kubernetes.io/projected/99387cea-c853-41ad-8926-39327379f170-kube-api-access-f62vb\") pod \"collect-profiles-29424525-4rxtv\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.493332 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.503506 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.507454 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q9qt\" (UniqueName: \"kubernetes.io/projected/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-kube-api-access-6q9qt\") pod \"cni-sysctl-allowlist-ds-nbqqf\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.510725 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod39e2e633_4eb3_4422_b9ef_5e81be4bff5e.slice/crio-4945dd1c306e4e701d61eae2cde0147841670d3199665cfc0142455ce72ae786 WatchSource:0}: Error finding container 4945dd1c306e4e701d61eae2cde0147841670d3199665cfc0142455ce72ae786: Status 404 returned error can't find the container with id 4945dd1c306e4e701d61eae2cde0147841670d3199665cfc0142455ce72ae786 Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.524522 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq86z\" (UniqueName: \"kubernetes.io/projected/584167c4-b6c8-4abb-80cd-7aaed553a463-kube-api-access-nq86z\") pod \"packageserver-7d4fc7d867-c2bmn\" (UID: \"584167c4-b6c8-4abb-80cd-7aaed553a463\") " pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.533918 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode52494b7_da4c_4aa1_82cc_285cc203d719.slice/crio-37f77bbbddf8c5c87a17471e7206047fa6d2661a70ea89379cbab0f4b171b08b WatchSource:0}: Error finding container 37f77bbbddf8c5c87a17471e7206047fa6d2661a70ea89379cbab0f4b171b08b: Status 404 returned error can't find the container with id 37f77bbbddf8c5c87a17471e7206047fa6d2661a70ea89379cbab0f4b171b08b Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.547631 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p89gb\" (UniqueName: \"kubernetes.io/projected/5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7-kube-api-access-p89gb\") pod \"control-plane-machine-set-operator-75ffdb6fcd-8cmcd\" (UID: \"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7\") " pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.552941 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-5777786469-52jsw"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.557084 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.565628 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-p58jx\" (UniqueName: \"kubernetes.io/projected/5860c4af-180f-4bda-a1cb-b4bb06c4f8e2-kube-api-access-p58jx\") pod \"service-ca-operator-5b9c976747-v57td\" (UID: \"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2\") " pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.572534 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.573233 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.073213101 +0000 UTC m=+124.752904567 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.584153 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" event={"ID":"36cbe175-1eb2-4280-9656-fb631e70056c","Type":"ContainerStarted","Data":"a5a71b9f44d2adb31493d52a5981af98c983c932e6e07227c6b02ae1ffe4505a"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.591504 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-66s5r\" (UniqueName: \"kubernetes.io/projected/2494549f-2d12-462f-9afd-c11172476e1e-kube-api-access-66s5r\") pod \"authentication-operator-7f5c659b84-8gf6m\" (UID: \"2494549f-2d12-462f-9afd-c11172476e1e\") " pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.592589 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" event={"ID":"0c6647e8-7cbf-4137-9bb6-2897bd47844c","Type":"ContainerStarted","Data":"929a508bc6c0c5def746ff25df364318757848fa12181a055c0781d69dcd624e"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.592635 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" event={"ID":"0c6647e8-7cbf-4137-9bb6-2897bd47844c","Type":"ContainerStarted","Data":"512770fc5c57d85549f8300bf06ead95aa7b30135d0a05795769e8610ea3dc3b"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.599548 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mxgl\" (UniqueName: \"kubernetes.io/projected/b125ace0-604e-48e1-808b-b7331df1c82a-kube-api-access-4mxgl\") pod \"machine-config-controller-f9cdd68f7-xzb7d\" (UID: \"b125ace0-604e-48e1-808b-b7331df1c82a\") " pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.606228 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.612820 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" event={"ID":"357b32d1-3358-408f-915c-bc92802a28b9","Type":"ContainerStarted","Data":"52ae5b425fa3888a13e3c10dad7da4743b8da104e16444906bf76feaa0013d9e"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.621333 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-747b44746d-bf7pr" event={"ID":"ce9e44ba-a2e2-443b-b265-aa312d55d7ca","Type":"ContainerStarted","Data":"e79a89339a3a2f16230b4b2da703dd1d7383b66682a4d768ab1680406e11cda0"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.621395 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-747b44746d-bf7pr" event={"ID":"ce9e44ba-a2e2-443b-b265-aa312d55d7ca","Type":"ContainerStarted","Data":"71a78348787c9cc57edf314dee82a6e89ea1545164870f0c9714df6d9cf40534"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.623011 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-799b87ffcd-ws7p5"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.630162 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" event={"ID":"4c398d16-ab6f-42b1-8f79-792bd8d2b637","Type":"ContainerStarted","Data":"d4a8665f76f26ce1c98b17fb16ac0d92aa3bbd58757e4e323d456bbc04e3dfc8"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.630448 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-x86zc\" (UniqueName: \"kubernetes.io/projected/c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd-kube-api-access-x86zc\") pod \"service-ca-74545575db-8cq9r\" (UID: \"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd\") " pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.636021 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4c2cr" event={"ID":"39e2e633-4eb3-4422-b9ef-5e81be4bff5e","Type":"ContainerStarted","Data":"4945dd1c306e4e701d61eae2cde0147841670d3199665cfc0142455ce72ae786"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.637511 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" event={"ID":"e52494b7-da4c-4aa1-82cc-285cc203d719","Type":"ContainerStarted","Data":"37f77bbbddf8c5c87a17471e7206047fa6d2661a70ea89379cbab0f4b171b08b"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.638622 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-sjqx5" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.639585 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/downloads-747b44746d-bf7pr" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.644019 5109 request.go:752] "Waited before sending request" delay="2.641578856s" reason="client-side throttling, not priority and fairness" verb="POST" URL="https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.646865 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d44f6ddf-hvlqh" event={"ID":"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d","Type":"ContainerStarted","Data":"cdab104672ec63ae7453f0b5f21868f40d0b02ee23af47f64cd1d42c9be517da"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.647806 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.647824 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh8xp\" (UniqueName: \"kubernetes.io/projected/68d5a689-abae-41b3-8c8b-0a9255fc276c-kube-api-access-dh8xp\") pod \"csi-hostpathplugin-pp75t\" (UID: \"68d5a689-abae-41b3-8c8b-0a9255fc276c\") " pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.653258 5109 patch_prober.go:28] interesting pod/downloads-747b44746d-bf7pr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.653313 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-bf7pr" podUID="ce9e44ba-a2e2-443b-b265-aa312d55d7ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.659288 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.678702 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.679066 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.179054105 +0000 UTC m=+124.858745571 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.679154 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.682372 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.689449 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" event={"ID":"ae8a431b-27bd-471a-b9d7-86211fc52f44","Type":"ContainerStarted","Data":"0558fdaaf8e1e9ef0ea2936c10d893812800c5c2db0034dc6ca1f2205c44d9c9"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.689532 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f430caa3-dc3a-4ca1-8246-7606ef18a297-kube-api-access\") pod \"openshift-kube-scheduler-operator-54f497555d-phm4k\" (UID: \"f430caa3-dc3a-4ca1-8246-7606ef18a297\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.690501 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.692062 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.693263 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhdwz\" (UniqueName: \"kubernetes.io/projected/0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e-kube-api-access-mhdwz\") pod \"openshift-apiserver-operator-846cbfc458-gqxz9\" (UID: \"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.701357 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" event={"ID":"0562c782-fba6-47a1-87a5-69422d01de09","Type":"ContainerStarted","Data":"d37c0050d22d55f01a043756d7a1bb39b7c25330a7dba4bdcdb49e38f283273b"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.704630 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-7xzzc\" (UniqueName: \"kubernetes.io/projected/693595ff-e600-48ce-9d44-2f996cc27307-kube-api-access-7xzzc\") pod \"oauth-openshift-66458b6674-mr25z\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.715659 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.718958 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.726072 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" event={"ID":"2d87a27d-ab14-48d9-a685-c43beae0f998","Type":"ContainerStarted","Data":"c41cf326a082dc91b4e1a1e9a077d67d789c92d8f072d6b67c618f4a0cde5697"} Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.726946 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlkvm\" (UniqueName: \"kubernetes.io/projected/6b4042ae-cae4-4d16-a997-9c236defaa8e-kube-api-access-hlkvm\") pod \"package-server-manager-77f986bd66-7kbk8\" (UID: \"6b4042ae-cae4-4d16-a997-9c236defaa8e\") " pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.728533 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.739321 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.744873 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.760203 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-td7d8\" (UniqueName: \"kubernetes.io/projected/68e67f52-4e97-4950-8678-575d8fa94f99-kube-api-access-td7d8\") pod \"olm-operator-5cdf44d969-qd2jd\" (UID: \"68e67f52-4e97-4950-8678-575d8fa94f99\") " pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.778883 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.779727 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.779960 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.27993252 +0000 UTC m=+124.959624026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.780382 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.780649 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.280642108 +0000 UTC m=+124.960333574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.781966 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.784417 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-69db94689b-sqnwm"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.787575 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-77chb"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.798034 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.805404 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.818909 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34a6df66_b508_4a7a_bc7e_254de76eaba7.slice/crio-b041e02b964f0e5b7e8fe3da1bc78a4b42a0873fbda54869042137fe867aed56 WatchSource:0}: Error finding container b041e02b964f0e5b7e8fe3da1bc78a4b42a0873fbda54869042137fe867aed56: Status 404 returned error can't find the container with id b041e02b964f0e5b7e8fe3da1bc78a4b42a0873fbda54869042137fe867aed56 Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.875489 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-74545575db-8cq9r" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.883028 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.883550 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.884169 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.384147327 +0000 UTC m=+125.063838803 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.886595 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" Dec 11 16:54:46 crc kubenswrapper[5109]: W1211 16:54:46.895112 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17170d9a_42cd_4a85_bafb_b6ced9f3503c.slice/crio-62260f6e13ab696550dd82a5055d84ef508c05dd177761c440564612558971fe WatchSource:0}: Error finding container 62260f6e13ab696550dd82a5055d84ef508c05dd177761c440564612558971fe: Status 404 returned error can't find the container with id 62260f6e13ab696550dd82a5055d84ef508c05dd177761c440564612558971fe Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.939705 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.957381 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm"] Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.965824 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.978491 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" Dec 11 16:54:46 crc kubenswrapper[5109]: I1211 16:54:46.985533 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:46 crc kubenswrapper[5109]: E1211 16:54:46.988485 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.488466784 +0000 UTC m=+125.168158250 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.050882 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.075014 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-747b44746d-bf7pr" podStartSLOduration=104.074996234 podStartE2EDuration="1m44.074996234s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:47.034510256 +0000 UTC m=+124.714201722" watchObservedRunningTime="2025-12-11 16:54:47.074996234 +0000 UTC m=+124.754687700" Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.087291 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.087873 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.587856224 +0000 UTC m=+125.267547690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.191555 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.194201 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.694182321 +0000 UTC m=+125.373873787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.280128 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-tzg7t"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.293093 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.293459 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.793443267 +0000 UTC m=+125.473134733 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.395208 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.395522 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.895509611 +0000 UTC m=+125.575201077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.499356 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.499655 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:47.999624684 +0000 UTC m=+125.679316150 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.500919 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.501451 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.001438079 +0000 UTC m=+125.681129545 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.528861 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.539926 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-sjqx5"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.541807 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-pp75t"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.546904 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.555812 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.595103 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.603662 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.604065 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.104048085 +0000 UTC m=+125.783739551 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: W1211 16:54:47.632125 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd7d2f41c_6e62_4dba_b44a_982d0ce9e256.slice/crio-89672f62e645f4e6836db7e0dfa275a7acfb7519091a6eb709fd4b1ed4dadc30 WatchSource:0}: Error finding container 89672f62e645f4e6836db7e0dfa275a7acfb7519091a6eb709fd4b1ed4dadc30: Status 404 returned error can't find the container with id 89672f62e645f4e6836db7e0dfa275a7acfb7519091a6eb709fd4b1ed4dadc30 Dec 11 16:54:47 crc kubenswrapper[5109]: W1211 16:54:47.673483 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99387cea_c853_41ad_8926_39327379f170.slice/crio-1bb468f0191664e88328af6ba5be9ffc64e4262df5ff7236b4c3737526b0205d WatchSource:0}: Error finding container 1bb468f0191664e88328af6ba5be9ffc64e4262df5ff7236b4c3737526b0205d: Status 404 returned error can't find the container with id 1bb468f0191664e88328af6ba5be9ffc64e4262df5ff7236b4c3737526b0205d Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.705505 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.706248 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.206201211 +0000 UTC m=+125.885892677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.724318 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.729403 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.757436 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.758271 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.773123 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.789830 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.791098 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" event={"ID":"1b24b31f-a9ae-4e65-88a9-6825ddff668e","Type":"ContainerStarted","Data":"9cf9855d0a42388cf0356c27d02df609dc261758326f4852b7dde8d930de1f38"} Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.795712 5109 generic.go:358] "Generic (PLEG): container finished" podID="ae8a431b-27bd-471a-b9d7-86211fc52f44" containerID="a2cc15ccd8480fcc644e81ea71db8fb023ad9b39762e3d415701782457e7bea3" exitCode=0 Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.795864 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" event={"ID":"ae8a431b-27bd-471a-b9d7-86211fc52f44","Type":"ContainerDied","Data":"a2cc15ccd8480fcc644e81ea71db8fb023ad9b39762e3d415701782457e7bea3"} Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.807575 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.807893 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.307877276 +0000 UTC m=+125.987568742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.809224 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" event={"ID":"0562c782-fba6-47a1-87a5-69422d01de09","Type":"ContainerStarted","Data":"007fa7ddb8a9da0946d560299de9ccafbe9f3a895fc29cc237c937ca0cc9a574"} Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.812914 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.814419 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-67c89758df-8rnw5"] Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.908947 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:47 crc kubenswrapper[5109]: E1211 16:54:47.909822 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.409811226 +0000 UTC m=+126.089502692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:47 crc kubenswrapper[5109]: I1211 16:54:47.924106 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-74545575db-8cq9r"] Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.007587 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d"] Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.011173 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.011416 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.511399598 +0000 UTC m=+126.191091064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.051442 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd"] Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.079201 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" event={"ID":"357b32d1-3358-408f-915c-bc92802a28b9","Type":"ContainerStarted","Data":"ac126cb790bc8aaa738d3e2358bed4690368ec870770a267175a99ac40f20636"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.087651 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" event={"ID":"99387cea-c853-41ad-8926-39327379f170","Type":"ContainerStarted","Data":"1bb468f0191664e88328af6ba5be9ffc64e4262df5ff7236b4c3737526b0205d"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.090650 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" event={"ID":"2e862e11-c81e-4697-bee2-8cb2b0ba590e","Type":"ContainerStarted","Data":"8995efcdee62a1315a2360725345c1ec45d1b4ba0a63ca7f49a6fcdd9074aca5"} Dec 11 16:54:48 crc kubenswrapper[5109]: W1211 16:54:48.106865 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68e67f52_4e97_4950_8678_575d8fa94f99.slice/crio-0314c11ce754723f1239c384d10a87aae1fa935d36846f9d22ce6c4acb49f9bf WatchSource:0}: Error finding container 0314c11ce754723f1239c384d10a87aae1fa935d36846f9d22ce6c4acb49f9bf: Status 404 returned error can't find the container with id 0314c11ce754723f1239c384d10a87aae1fa935d36846f9d22ce6c4acb49f9bf Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.107950 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" event={"ID":"68d5a689-abae-41b3-8c8b-0a9255fc276c","Type":"ContainerStarted","Data":"f0c54bacf7956073dcfe99a27872d1d6da304574ffdc7db1499d3ee80e553849"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.112239 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.112599 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.612583881 +0000 UTC m=+126.292275367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.173426 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-sjqx5" event={"ID":"d7d2f41c-6e62-4dba-b44a-982d0ce9e256","Type":"ContainerStarted","Data":"89672f62e645f4e6836db7e0dfa275a7acfb7519091a6eb709fd4b1ed4dadc30"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.179570 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" event={"ID":"28a5c354-8565-49aa-b329-cf529c594431","Type":"ContainerStarted","Data":"4f7462522b8a42127e1635bd5700c1e24e9c3cd31ffb9d71a4a78d23501d402d"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.193788 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-77chb" event={"ID":"17170d9a-42cd-4a85-bafb-b6ced9f3503c","Type":"ContainerStarted","Data":"62260f6e13ab696550dd82a5055d84ef508c05dd177761c440564612558971fe"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.194699 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" podStartSLOduration=105.194652793 podStartE2EDuration="1m45.194652793s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:48.159702659 +0000 UTC m=+125.839394145" watchObservedRunningTime="2025-12-11 16:54:48.194652793 +0000 UTC m=+125.874344259" Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.215361 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.215723 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.715705411 +0000 UTC m=+126.395396877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.217203 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" event={"ID":"4cb7d92e-f9a5-4deb-a375-7f48da3f0899","Type":"ContainerStarted","Data":"b22632c7a6643add133b69d6ca523c059ebdf205547692ef3a2b0c2daeb3f695"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.267517 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mr25z"] Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.317626 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.318103 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.818091652 +0000 UTC m=+126.497783118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.323795 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" event={"ID":"4eed92a0-6f24-424d-9be1-1f7e336ce8a9","Type":"ContainerStarted","Data":"c51d5399e4312d80df519f40cafa445d5f6ff8237b195392a2c76bab68f34e41"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.327768 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" event={"ID":"aa291f60-d16f-476b-a0b0-0e988aa54e3b","Type":"ContainerStarted","Data":"2f3cc549dfd7840794243b84e079ae8c58cf6e6e22a2698a66fff5ae6db45c22"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.347046 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-4c2cr" event={"ID":"39e2e633-4eb3-4422-b9ef-5e81be4bff5e","Type":"ContainerStarted","Data":"f09a10b9e517ed5412dd3c85f35bf688c2cbe119f6eb80fcb9dc2e9abee70ced"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.358718 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" event={"ID":"eca35b32-bf7f-49d0-b796-060590d7c233","Type":"ContainerStarted","Data":"eba09091a68df55d7756daf6a4ffaa3e64bfd782aa731613d971efd62c1b38f5"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.360700 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" event={"ID":"637dcb89-3251-4807-8bf0-3c844ca1c05a","Type":"ContainerStarted","Data":"d8edda99621637d34b34adbf6e71d300503db4b45678fde08158cd60cabd8876"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.362111 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" event={"ID":"34a6df66-b508-4a7a-bc7e-254de76eaba7","Type":"ContainerStarted","Data":"b041e02b964f0e5b7e8fe3da1bc78a4b42a0873fbda54869042137fe867aed56"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.368066 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-4c2cr" podStartSLOduration=6.368053998 podStartE2EDuration="6.368053998s" podCreationTimestamp="2025-12-11 16:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:48.367457034 +0000 UTC m=+126.047148510" watchObservedRunningTime="2025-12-11 16:54:48.368053998 +0000 UTC m=+126.047745465" Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.368815 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-64d44f6ddf-hvlqh" event={"ID":"e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d","Type":"ContainerStarted","Data":"6f392d436c86ccf2beba2f1cd666bde8c11e1b3ee151db54da20220180a4b57a"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.372723 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" event={"ID":"36cbe175-1eb2-4280-9656-fb631e70056c","Type":"ContainerStarted","Data":"f45c3b362cbb2102d3464395e918960faef7440a5478acb8273789e57c4a9798"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.376448 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" event={"ID":"0c6647e8-7cbf-4137-9bb6-2897bd47844c","Type":"ContainerStarted","Data":"5d8abaa8d92be84671c005c0077217cfe45d43ff22b0322e80caef94d6448cf6"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.386913 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" event={"ID":"099e943e-0a16-48e4-b19d-0cbf2dc39466","Type":"ContainerStarted","Data":"aa011e91ea777ee0fdcece341298e7cb8d2835e6ba0371b999ecabe16ef9f5df"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.392177 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" event={"ID":"e707a0e8-be77-43fb-8236-19159201399a","Type":"ContainerStarted","Data":"91aa9cdb3e6f08f48376383919f2a5b87c9c97bab3c974a47662a5e9a704f0f9"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.400534 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-64d44f6ddf-hvlqh" podStartSLOduration=105.400515382 podStartE2EDuration="1m45.400515382s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:48.396351372 +0000 UTC m=+126.076042838" watchObservedRunningTime="2025-12-11 16:54:48.400515382 +0000 UTC m=+126.080206848" Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.406870 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" event={"ID":"4c398d16-ab6f-42b1-8f79-792bd8d2b637","Type":"ContainerStarted","Data":"4d132747115d5620d6d33c5cb427b4463022b67824a5c68e8bbaef4e9f36038e"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.415527 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" event={"ID":"c805a7f3-2074-4cfa-9001-aee570677c8e","Type":"ContainerStarted","Data":"c8dc3e4182ae5160405c8cc8f525bbb4aa9d8e74d0325a9c0bba903120b2e77b"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.418582 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.419029 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:48.918985368 +0000 UTC m=+126.598676844 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.444671 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-686468bdd5-xr7js" podStartSLOduration=105.444654527 podStartE2EDuration="1m45.444654527s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:48.433769825 +0000 UTC m=+126.113461301" watchObservedRunningTime="2025-12-11 16:54:48.444654527 +0000 UTC m=+126.124345993" Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.452852 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" event={"ID":"455d2dca-4769-4478-bbe5-ba72ed2db6ba","Type":"ContainerStarted","Data":"8c7c774de3d589dbb60fba84c04180daf2f1fe42f4e284d186cf4650d7af38d7"} Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.458931 5109 patch_prober.go:28] interesting pod/downloads-747b44746d-bf7pr container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.458978 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-747b44746d-bf7pr" podUID="ce9e44ba-a2e2-443b-b265-aa312d55d7ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.491312 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-6b9cb4dbcf-zx6sp" podStartSLOduration=105.491270953 podStartE2EDuration="1m45.491270953s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:48.480721578 +0000 UTC m=+126.160413044" watchObservedRunningTime="2025-12-11 16:54:48.491270953 +0000 UTC m=+126.170962419" Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.517433 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" podStartSLOduration=105.517414223 podStartE2EDuration="1m45.517414223s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:48.517048865 +0000 UTC m=+126.196740331" watchObservedRunningTime="2025-12-11 16:54:48.517414223 +0000 UTC m=+126.197105689" Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.520304 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.521275 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.021255976 +0000 UTC m=+126.700947442 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.623636 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.624098 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.124076429 +0000 UTC m=+126.803767895 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.725935 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.726247 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.226233744 +0000 UTC m=+126.905925210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.828853 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.829137 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.329121988 +0000 UTC m=+127.008813454 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:48 crc kubenswrapper[5109]: I1211 16:54:48.931971 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:48 crc kubenswrapper[5109]: E1211 16:54:48.933086 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.433060757 +0000 UTC m=+127.112752223 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.033426 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.033897 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.533879761 +0000 UTC m=+127.213571227 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.134652 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.135265 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.635247089 +0000 UTC m=+127.314938555 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.236585 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.236815 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.736784359 +0000 UTC m=+127.416475825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.237357 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.237652 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.73763894 +0000 UTC m=+127.417330406 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.346849 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.347368 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.847347338 +0000 UTC m=+127.527038804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.450994 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.451533 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:49.951520544 +0000 UTC m=+127.631212010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.564044 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.564348 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.064332576 +0000 UTC m=+127.744024042 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.568617 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-77chb" event={"ID":"17170d9a-42cd-4a85-bafb-b6ced9f3503c","Type":"ContainerStarted","Data":"32cbbb8f68274243daaafc219ef308cc1515ac59053a5d86d16f8e31759c7a35"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.607840 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-sjqx5" event={"ID":"d7d2f41c-6e62-4dba-b44a-982d0ce9e256","Type":"ContainerStarted","Data":"1420fa5e51e4d25d10012af5b00873aba0aee1f2d0f629a4f50ed6a1848205ab"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.636346 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-sjqx5" podStartSLOduration=7.636330384 podStartE2EDuration="7.636330384s" podCreationTimestamp="2025-12-11 16:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:49.636010156 +0000 UTC m=+127.315701622" watchObservedRunningTime="2025-12-11 16:54:49.636330384 +0000 UTC m=+127.316021850" Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.637286 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" event={"ID":"28a5c354-8565-49aa-b329-cf529c594431","Type":"ContainerStarted","Data":"2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.638852 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.651725 5109 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-tzg7t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.651790 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.670312 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.670583 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.170571541 +0000 UTC m=+127.850263007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.675328 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" event={"ID":"eca35b32-bf7f-49d0-b796-060590d7c233","Type":"ContainerStarted","Data":"7dc52400c3991085760e79e8fa52be8ec5b91b8e9296e2c9889c6f6078b986c2"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.677668 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" event={"ID":"34a6df66-b508-4a7a-bc7e-254de76eaba7","Type":"ContainerStarted","Data":"68ce6eafcb3383adb27365c49dac2103d0d71e4cf5b7d5741a34578d7bfc8d8b"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.719159 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" podStartSLOduration=106.719142194 podStartE2EDuration="1m46.719142194s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:49.690363629 +0000 UTC m=+127.370055105" watchObservedRunningTime="2025-12-11 16:54:49.719142194 +0000 UTC m=+127.398833660" Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.766929 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-575994946d-crdzk" podStartSLOduration=106.766910607 podStartE2EDuration="1m46.766910607s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:49.73104083 +0000 UTC m=+127.410732296" watchObservedRunningTime="2025-12-11 16:54:49.766910607 +0000 UTC m=+127.446602073" Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.767523 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-69b85846b6-6jvnh" podStartSLOduration=106.767518871 podStartE2EDuration="1m46.767518871s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:49.767192773 +0000 UTC m=+127.446884229" watchObservedRunningTime="2025-12-11 16:54:49.767518871 +0000 UTC m=+127.447210327" Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.775136 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.776326 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.276310763 +0000 UTC m=+127.956002219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.791974 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" event={"ID":"e707a0e8-be77-43fb-8236-19159201399a","Type":"ContainerStarted","Data":"c992cb1c8dba96b84673f10ba908d2f52dc11f8a08133f9005a55f0978558f29"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.792030 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" event={"ID":"e707a0e8-be77-43fb-8236-19159201399a","Type":"ContainerStarted","Data":"3e9a2ba3abdc23f68039d4c28a6975c79c01628a725e3a33d218115b526f82cd"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.813381 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" event={"ID":"6b4042ae-cae4-4d16-a997-9c236defaa8e","Type":"ContainerStarted","Data":"52be2afe97f40bc6ea6b541bc959fb33851cbe844dfe4c1c40a2a571a2504c28"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.834446 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-67c9d58cbb-9hbcp" podStartSLOduration=106.834424766 podStartE2EDuration="1m46.834424766s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:49.831351882 +0000 UTC m=+127.511043348" watchObservedRunningTime="2025-12-11 16:54:49.834424766 +0000 UTC m=+127.514116242" Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.855109 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-565b79b866-l8msx" event={"ID":"455d2dca-4769-4478-bbe5-ba72ed2db6ba","Type":"ContainerStarted","Data":"bc96248b2d1db1e65be417a5e257bbb11a7e16b9cc2954b54fe9e37af7a30503"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.885166 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.885896 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.385882559 +0000 UTC m=+128.065574025 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.934896 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" event={"ID":"1b24b31f-a9ae-4e65-88a9-6825ddff668e","Type":"ContainerStarted","Data":"0cf461dae2727c033dcf8d2f775d328895b49f5da47b301d84d1b484a26cd2c9"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.950904 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" event={"ID":"2494549f-2d12-462f-9afd-c11172476e1e","Type":"ContainerStarted","Data":"18a4d227ec0501fba66444fdf8349368d49577d83de908462ad179b11e2a1bac"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.950949 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" event={"ID":"2494549f-2d12-462f-9afd-c11172476e1e","Type":"ContainerStarted","Data":"5046b12c35aa8940acf494cdd84cc3ac61e29df9a6b5cadd1fd67ac4c61b31c1"} Dec 11 16:54:49 crc kubenswrapper[5109]: I1211 16:54:49.988719 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:49 crc kubenswrapper[5109]: E1211 16:54:49.989031 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.488999637 +0000 UTC m=+128.168691103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:49.991993 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-7f5c659b84-8gf6m" podStartSLOduration=106.99197942 podStartE2EDuration="1m46.99197942s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:49.988549557 +0000 UTC m=+127.668241023" watchObservedRunningTime="2025-12-11 16:54:49.99197942 +0000 UTC m=+127.671670886" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.025437 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" event={"ID":"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7","Type":"ContainerStarted","Data":"b86059f04aa1a68c5928dc21a00fb92baef90264665222679ec63f2136bcab9d"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.026683 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" event={"ID":"0562c782-fba6-47a1-87a5-69422d01de09","Type":"ContainerStarted","Data":"e1483f02af8302b466f70c143d8d36f028a09480ab791095464704c73567412b"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.034111 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" event={"ID":"99387cea-c853-41ad-8926-39327379f170","Type":"ContainerStarted","Data":"c3a49be7c1286b87feb1651ae54c1a9acfc3aab9d04bc7012fedd61313a200a0"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.057315 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" event={"ID":"2e862e11-c81e-4697-bee2-8cb2b0ba590e","Type":"ContainerStarted","Data":"7318b9c1d752b6ed4054643a8c79476ca0a0b4e6997ad627c71df8b568f1906d"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.059656 5109 generic.go:358] "Generic (PLEG): container finished" podID="4cb7d92e-f9a5-4deb-a375-7f48da3f0899" containerID="53fa0773699aab1cbe8d5924ae4a42beceb12dcb19f74f19581fbe608e9bbe49" exitCode=0 Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.059766 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" event={"ID":"4cb7d92e-f9a5-4deb-a375-7f48da3f0899","Type":"ContainerDied","Data":"53fa0773699aab1cbe8d5924ae4a42beceb12dcb19f74f19581fbe608e9bbe49"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.073102 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" event={"ID":"b125ace0-604e-48e1-808b-b7331df1c82a","Type":"ContainerStarted","Data":"d12638f5bf2e1313928dc478c0c7277c0ade3de58e06cd6d3a58c66b5ae7958b"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.073151 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" event={"ID":"b125ace0-604e-48e1-808b-b7331df1c82a","Type":"ContainerStarted","Data":"be3d26bf308e65d3c1e7152524db12b5c009224ccbc9ca2651a3fdc31940ed86"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.089192 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" event={"ID":"6116425c-a3f6-4a09-91b1-207560b62779","Type":"ContainerStarted","Data":"df5112be7ec0f18e7d12c3e443385d2f88a327f90af112a02187727c4d89b275"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.089245 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" event={"ID":"6116425c-a3f6-4a09-91b1-207560b62779","Type":"ContainerStarted","Data":"075becd1407dd930d2b3d90c523aa2044171964a234f0e3140db949d4daec39a"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.090211 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.091348 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.591336027 +0000 UTC m=+128.271027493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.097492 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" event={"ID":"4eed92a0-6f24-424d-9be1-1f7e336ce8a9","Type":"ContainerStarted","Data":"b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.099399 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" event={"ID":"aa291f60-d16f-476b-a0b0-0e988aa54e3b","Type":"ContainerStarted","Data":"c59e3324375dc19ee4c563d1b53411dd9c6a6b416267755e65b521566d2d89cb"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.100632 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" event={"ID":"73f65af5-9773-4165-8621-bf20d9eefd27","Type":"ContainerStarted","Data":"a69aac2c6bc9eb835cc0b64d3e0b866c2d22d7cabf42344497ace56c3bd3536f"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.107366 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" event={"ID":"637dcb89-3251-4807-8bf0-3c844ca1c05a","Type":"ContainerStarted","Data":"bb89f803fabe7d1beec96a870c93417961d5d3d67d3230ebccb4df9456464e8e"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.122199 5109 generic.go:358] "Generic (PLEG): container finished" podID="2d87a27d-ab14-48d9-a685-c43beae0f998" containerID="495f7bd14d6b15c34449f7a966504d3d617a320e8963ec509fd568113d6b1c2b" exitCode=0 Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.123171 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" event={"ID":"2d87a27d-ab14-48d9-a685-c43beae0f998","Type":"ContainerDied","Data":"495f7bd14d6b15c34449f7a966504d3d617a320e8963ec509fd568113d6b1c2b"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.144522 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.145150 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" event={"ID":"099e943e-0a16-48e4-b19d-0cbf2dc39466","Type":"ContainerStarted","Data":"c3000258e9bf6e466fb95a964fde0ffd6ab0663f1800f4cd02ef8ce3fc0d90cd"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.154902 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" event={"ID":"693595ff-e600-48ce-9d44-2f996cc27307","Type":"ContainerStarted","Data":"308e6159c7ca96fe7a06acf3961378c3d2ad623326b8d34b39037b73685f9a65"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.173940 5109 patch_prober.go:28] interesting pod/console-operator-67c89758df-8rnw5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.174001 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" podUID="6116425c-a3f6-4a09-91b1-207560b62779" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/readyz\": dial tcp 10.217.0.26:8443: connect: connection refused" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.180872 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" podStartSLOduration=108.180852018 podStartE2EDuration="1m48.180852018s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.180593222 +0000 UTC m=+127.860284688" watchObservedRunningTime="2025-12-11 16:54:50.180852018 +0000 UTC m=+127.860543484" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.193471 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.194726 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.694707243 +0000 UTC m=+128.374398709 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.196340 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" event={"ID":"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e","Type":"ContainerStarted","Data":"3eb22b167d6b9e8cd63f1a27fbf483fe5beb8f84980f184b86da84671f8828aa"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.228057 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" event={"ID":"4c398d16-ab6f-42b1-8f79-792bd8d2b637","Type":"ContainerStarted","Data":"efda87b204df1ab16d4c0e3f1e79917fac17f2a585801ea7575bdb97cce06085"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.239871 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podStartSLOduration=107.239848093 podStartE2EDuration="1m47.239848093s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.213884806 +0000 UTC m=+127.893576272" watchObservedRunningTime="2025-12-11 16:54:50.239848093 +0000 UTC m=+127.919539559" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.243180 5109 ???:1] "http: TLS handshake error from 192.168.126.11:32980: no serving certificate available for the kubelet" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.247139 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" event={"ID":"c805a7f3-2074-4cfa-9001-aee570677c8e","Type":"ContainerStarted","Data":"d1791cf0c1ae9f919f9f8cf12289416858708f8fa3e8ad895035d59ecb835780"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.252850 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" event={"ID":"68e67f52-4e97-4950-8678-575d8fa94f99","Type":"ContainerStarted","Data":"e7030c9dbdce2fec70849ce50a10a9202bc3d35c03406a3cc5315a39bbd13ea6"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.252899 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" event={"ID":"68e67f52-4e97-4950-8678-575d8fa94f99","Type":"ContainerStarted","Data":"0314c11ce754723f1239c384d10a87aae1fa935d36846f9d22ce6c4acb49f9bf"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.256228 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" event={"ID":"f430caa3-dc3a-4ca1-8246-7606ef18a297","Type":"ContainerStarted","Data":"b1268999e83e06bd85d2cccf9385b5ef5ac324770f448362d992c7889d8ac654"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.257637 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" event={"ID":"e52494b7-da4c-4aa1-82cc-285cc203d719","Type":"ContainerStarted","Data":"e752a9ba4d3d1b106f525cb419736f25a555d5005f334a1ba848fa3db90f3d1b"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.258528 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-74545575db-8cq9r" event={"ID":"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd","Type":"ContainerStarted","Data":"71c8ffd97c72e9f09bbf32ecfae71eec7c79f96d1f9e414723d3932b70489ddf"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.263290 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-54c688565-5lz5m" podStartSLOduration=108.263272808 podStartE2EDuration="1m48.263272808s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.263183916 +0000 UTC m=+127.942875382" watchObservedRunningTime="2025-12-11 16:54:50.263272808 +0000 UTC m=+127.942964274" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.263414 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" podStartSLOduration=107.263409901 podStartE2EDuration="1m47.263409901s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.24259922 +0000 UTC m=+127.922290686" watchObservedRunningTime="2025-12-11 16:54:50.263409901 +0000 UTC m=+127.943101367" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.266338 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" event={"ID":"584167c4-b6c8-4abb-80cd-7aaed553a463","Type":"ContainerStarted","Data":"40e834b3e8b91521fcfc4c9eccb0bbcd0a973474d2c805965dbc8ead436e4cd4"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.296613 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.302235 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.802222118 +0000 UTC m=+128.481913584 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.332941 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" event={"ID":"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2","Type":"ContainerStarted","Data":"922c41e13c710130a0f96e849690861fc61ed7a3a75f80f1e83698180faf0474"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.339111 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" event={"ID":"357b32d1-3358-408f-915c-bc92802a28b9","Type":"ContainerStarted","Data":"4a0aaeb7997f58fdc2e2291b21c8d71c7c6b82304359309e7689cba66facff70"} Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.343250 5109 ???:1] "http: TLS handshake error from 192.168.126.11:32992: no serving certificate available for the kubelet" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.397712 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.397874 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.897847436 +0000 UTC m=+128.577538902 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.398181 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.398509 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.898495643 +0000 UTC m=+128.578187109 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.429562 5109 ???:1] "http: TLS handshake error from 192.168.126.11:33002: no serving certificate available for the kubelet" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.494480 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.494514 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.498890 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.499035 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.999004069 +0000 UTC m=+128.678695535 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.499507 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.499889 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:50.999877019 +0000 UTC m=+128.679568475 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.500716 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.516844 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" podStartSLOduration=107.516819238 podStartE2EDuration="1m47.516819238s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.315062698 +0000 UTC m=+127.994754164" watchObservedRunningTime="2025-12-11 16:54:50.516819238 +0000 UTC m=+128.196510704" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.518855 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" podStartSLOduration=8.518844918 podStartE2EDuration="8.518844918s" podCreationTimestamp="2025-12-11 16:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.516255865 +0000 UTC m=+128.195947331" watchObservedRunningTime="2025-12-11 16:54:50.518844918 +0000 UTC m=+128.198536384" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.524449 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.526101 5109 ???:1] "http: TLS handshake error from 192.168.126.11:33016: no serving certificate available for the kubelet" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.531163 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-5cdf44d969-qd2jd" podStartSLOduration=107.531131115 podStartE2EDuration="1m47.531131115s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.529808253 +0000 UTC m=+128.209499729" watchObservedRunningTime="2025-12-11 16:54:50.531131115 +0000 UTC m=+128.210822591" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.552052 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-86c45576b9-njh8k" podStartSLOduration=107.552034269 podStartE2EDuration="1m47.552034269s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.549138928 +0000 UTC m=+128.228830384" watchObservedRunningTime="2025-12-11 16:54:50.552034269 +0000 UTC m=+128.231725725" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.578620 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" podStartSLOduration=107.57860334 podStartE2EDuration="1m47.57860334s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.57693414 +0000 UTC m=+128.256625606" watchObservedRunningTime="2025-12-11 16:54:50.57860334 +0000 UTC m=+128.258294796" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.600627 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.600778 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.100754945 +0000 UTC m=+128.780446411 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.601423 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.602312 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.102297442 +0000 UTC m=+128.781988908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.606711 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.609812 5109 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-lth2t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 16:54:50 crc kubenswrapper[5109]: [-]has-synced failed: reason withheld Dec 11 16:54:50 crc kubenswrapper[5109]: [+]process-running ok Dec 11 16:54:50 crc kubenswrapper[5109]: healthz check failed Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.609920 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podUID="2e862e11-c81e-4697-bee2-8cb2b0ba590e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.641106 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55658: no serving certificate available for the kubelet" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.669531 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-69d5f845f8-lcqcm" podStartSLOduration=107.669511514 podStartE2EDuration="1m47.669511514s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.668418339 +0000 UTC m=+128.348109825" watchObservedRunningTime="2025-12-11 16:54:50.669511514 +0000 UTC m=+128.349202980" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.702961 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.703951 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.203934766 +0000 UTC m=+128.883626232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.737483 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" podStartSLOduration=107.737462155 podStartE2EDuration="1m47.737462155s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.700537974 +0000 UTC m=+128.380229450" watchObservedRunningTime="2025-12-11 16:54:50.737462155 +0000 UTC m=+128.417153651" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.737971 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" podStartSLOduration=107.737964197 podStartE2EDuration="1m47.737964197s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.73642031 +0000 UTC m=+128.416111776" watchObservedRunningTime="2025-12-11 16:54:50.737964197 +0000 UTC m=+128.417655663" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.745587 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55672: no serving certificate available for the kubelet" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.766087 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-74545575db-8cq9r" podStartSLOduration=107.766069865 podStartE2EDuration="1m47.766069865s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.765202775 +0000 UTC m=+128.444894241" watchObservedRunningTime="2025-12-11 16:54:50.766069865 +0000 UTC m=+128.445761351" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.786249 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-6b564684c8-84ffz" podStartSLOduration=108.786228052 podStartE2EDuration="1m48.786228052s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.783676681 +0000 UTC m=+128.463368147" watchObservedRunningTime="2025-12-11 16:54:50.786228052 +0000 UTC m=+128.465919518" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.803655 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-755bb95488-dgfs6" podStartSLOduration=107.803639212 podStartE2EDuration="1m47.803639212s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.801925521 +0000 UTC m=+128.481616987" watchObservedRunningTime="2025-12-11 16:54:50.803639212 +0000 UTC m=+128.483330678" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.806086 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.806381 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.306368648 +0000 UTC m=+128.986060114 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.824137 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" podStartSLOduration=107.824118077 podStartE2EDuration="1m47.824118077s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.823408479 +0000 UTC m=+128.503099945" watchObservedRunningTime="2025-12-11 16:54:50.824118077 +0000 UTC m=+128.503809543" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.847904 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55680: no serving certificate available for the kubelet" Dec 11 16:54:50 crc kubenswrapper[5109]: I1211 16:54:50.910298 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:50 crc kubenswrapper[5109]: E1211 16:54:50.910669 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.410651906 +0000 UTC m=+129.090343372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.012067 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.012491 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.512474323 +0000 UTC m=+129.192165789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.035478 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55686: no serving certificate available for the kubelet" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.113643 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.113867 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.613834851 +0000 UTC m=+129.293526317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.154100 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" podStartSLOduration=108.154080502 podStartE2EDuration="1m48.154080502s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:50.857205045 +0000 UTC m=+128.536896511" watchObservedRunningTime="2025-12-11 16:54:51.154080502 +0000 UTC m=+128.833771968" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.155330 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-nbqqf"] Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.215646 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.216101 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.716083659 +0000 UTC m=+129.395775125 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.317037 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.317197 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.817169359 +0000 UTC m=+129.496860825 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.317328 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.317650 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.81763764 +0000 UTC m=+129.497329106 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.357141 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" event={"ID":"2d87a27d-ab14-48d9-a685-c43beae0f998","Type":"ContainerStarted","Data":"b87f4a1bbb72362a3d4369fe28e923acf7fc1cb0467180e7a7ea4aa4b619ad00"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.378214 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" event={"ID":"693595ff-e600-48ce-9d44-2f996cc27307","Type":"ContainerStarted","Data":"e57f87ee47ac6a19a1505c54775ea70c7f2ef80da763216df55ef895048d5412"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.378563 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.385405 5109 patch_prober.go:28] interesting pod/oauth-openshift-66458b6674-mr25z container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.31:6443/healthz\": dial tcp 10.217.0.31:6443: connect: connection refused" start-of-body= Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.385513 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" podUID="693595ff-e600-48ce-9d44-2f996cc27307" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.31:6443/healthz\": dial tcp 10.217.0.31:6443: connect: connection refused" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.405798 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-846cbfc458-gqxz9" event={"ID":"0cc8d01a-e7fb-4c68-91cb-f13f08d0a28e","Type":"ContainerStarted","Data":"af52467f150fa029e6ba7124ff75a865ac94fc6a499d8298501a7b7dc7dd53ae"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.420962 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.421461 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:51.921444275 +0000 UTC m=+129.601135741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.442650 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" event={"ID":"f430caa3-dc3a-4ca1-8246-7606ef18a297","Type":"ContainerStarted","Data":"892bce8aae3e874671c31d5f74bec4394f46726f1ea7f8341e4adfb45c3c4273"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.459594 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-74545575db-8cq9r" event={"ID":"c29028a0-9e7b-4c19-b1ed-b4189c5fe1cd","Type":"ContainerStarted","Data":"67848c037adb904bf1ffbb7011a1032cb40422e6fd4b113516d222bb787e32b8"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.464340 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" event={"ID":"584167c4-b6c8-4abb-80cd-7aaed553a463","Type":"ContainerStarted","Data":"ae21feab8a970fbfc3ab0d08737ed22c4a5b7ade8e4b5191a1f498d0c6e1b023"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.465210 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.474104 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" podStartSLOduration=108.474089697 podStartE2EDuration="1m48.474089697s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.420182596 +0000 UTC m=+129.099874062" watchObservedRunningTime="2025-12-11 16:54:51.474089697 +0000 UTC m=+129.153781163" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.491684 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-5b9c976747-v57td" event={"ID":"5860c4af-180f-4bda-a1cb-b4bb06c4f8e2","Type":"ContainerStarted","Data":"b53d513df5fb62b6c89f2ebb191f97b2f1f7ac574d305aea3c2a64d9a64ce4e4"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.525759 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.526286 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.026269136 +0000 UTC m=+129.705960672 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.534161 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" event={"ID":"ae8a431b-27bd-471a-b9d7-86211fc52f44","Type":"ContainerStarted","Data":"9030b47e8b97e047f7a32beacac3a2a63cd815b4a291011c0c41a67b49db2d34"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.555669 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-77chb" event={"ID":"17170d9a-42cd-4a85-bafb-b6ced9f3503c","Type":"ContainerStarted","Data":"b2724dfafaa15a0194d4cc0a1a6563cebfc0eef5cae6866b1fd5cfc5f6b23e25"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.556613 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-dns/dns-default-77chb" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.561838 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-54f497555d-phm4k" podStartSLOduration=108.561821305 podStartE2EDuration="1m48.561821305s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.474408954 +0000 UTC m=+129.154100420" watchObservedRunningTime="2025-12-11 16:54:51.561821305 +0000 UTC m=+129.241512771" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.568655 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" event={"ID":"6b4042ae-cae4-4d16-a997-9c236defaa8e","Type":"ContainerStarted","Data":"b99a9431085788c09e0c4b72e9ccaf61f37457b36d41c2d45ffabe22fc81740d"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.568692 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" event={"ID":"6b4042ae-cae4-4d16-a997-9c236defaa8e","Type":"ContainerStarted","Data":"6177da824d796ef696c2dfb197faeeccdd4839f317bdd383cbb09f58e3a39308"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.569421 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.572099 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" event={"ID":"1b24b31f-a9ae-4e65-88a9-6825ddff668e","Type":"ContainerStarted","Data":"6821416c8ca14babb5caa505b2c7bedcf9bc099953294ae7e460f6ee446d3ea7"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.591065 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-75ffdb6fcd-8cmcd" event={"ID":"5c2d3c52-0a65-4f5d-ab5c-fdc9178fbfd7","Type":"ContainerStarted","Data":"8cf568cd1d01c51c58d1857d9cc20b63cc76338a061fec5c5bc465cd34a72629"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.617252 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" podStartSLOduration=108.617235272 podStartE2EDuration="1m48.617235272s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.564411788 +0000 UTC m=+129.244103254" watchObservedRunningTime="2025-12-11 16:54:51.617235272 +0000 UTC m=+129.296926738" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.617775 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-77chb" podStartSLOduration=9.617771326 podStartE2EDuration="9.617771326s" podCreationTimestamp="2025-12-11 16:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.615377438 +0000 UTC m=+129.295068904" watchObservedRunningTime="2025-12-11 16:54:51.617771326 +0000 UTC m=+129.297462782" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.625810 5109 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-lth2t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 16:54:51 crc kubenswrapper[5109]: [-]has-synced failed: reason withheld Dec 11 16:54:51 crc kubenswrapper[5109]: [+]process-running ok Dec 11 16:54:51 crc kubenswrapper[5109]: healthz check failed Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.625878 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podUID="2e862e11-c81e-4697-bee2-8cb2b0ba590e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.627392 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.628654 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.128637677 +0000 UTC m=+129.808329143 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.630347 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" event={"ID":"4cb7d92e-f9a5-4deb-a375-7f48da3f0899","Type":"ContainerStarted","Data":"a1716f45ed3bc781e0a5eb84007611e7b58fd1832dfbb0c9006a28c220467038"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.630768 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.648334 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-69db94689b-sqnwm" podStartSLOduration=108.648317832 podStartE2EDuration="1m48.648317832s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.647237147 +0000 UTC m=+129.326928613" watchObservedRunningTime="2025-12-11 16:54:51.648317832 +0000 UTC m=+129.328009298" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.661606 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" event={"ID":"b125ace0-604e-48e1-808b-b7331df1c82a","Type":"ContainerStarted","Data":"6abe70f2daee1dabb0a7dc8087133ffe69c3a54fed7cbe8663a3561baa487321"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.691999 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" podStartSLOduration=108.691978897 podStartE2EDuration="1m48.691978897s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.691051894 +0000 UTC m=+129.370743400" watchObservedRunningTime="2025-12-11 16:54:51.691978897 +0000 UTC m=+129.371670363" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.695026 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" event={"ID":"aa291f60-d16f-476b-a0b0-0e988aa54e3b","Type":"ContainerStarted","Data":"5195ab980cc2b2cd67b027478b3d8d03541f219033e890b142294e8895ca7aae"} Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.714444 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v6ctk"] Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.729667 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.730303 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.230292412 +0000 UTC m=+129.909983868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.743104 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-f9cdd68f7-xzb7d" podStartSLOduration=108.74308297 podStartE2EDuration="1m48.74308297s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.724907011 +0000 UTC m=+129.404598477" watchObservedRunningTime="2025-12-11 16:54:51.74308297 +0000 UTC m=+129.422774456" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.756328 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" podStartSLOduration=109.75631386 podStartE2EDuration="1m49.75631386s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.755157482 +0000 UTC m=+129.434848938" watchObservedRunningTime="2025-12-11 16:54:51.75631386 +0000 UTC m=+129.436005326" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.763506 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55694: no serving certificate available for the kubelet" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.832241 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.832590 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.33255887 +0000 UTC m=+130.012250336 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.890622 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-799b87ffcd-ws7p5" podStartSLOduration=108.890599462 podStartE2EDuration="1m48.890599462s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:51.782875261 +0000 UTC m=+129.462566727" watchObservedRunningTime="2025-12-11 16:54:51.890599462 +0000 UTC m=+129.570290928" Dec 11 16:54:51 crc kubenswrapper[5109]: I1211 16:54:51.934137 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:51 crc kubenswrapper[5109]: E1211 16:54:51.934502 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.434484571 +0000 UTC m=+130.114176037 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.035109 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.035277 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.535244183 +0000 UTC m=+130.214935649 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.035658 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.035987 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.53597468 +0000 UTC m=+130.215666146 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.136358 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.136570 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.636540038 +0000 UTC m=+130.316231504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.136653 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.136975 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.636962358 +0000 UTC m=+130.316653824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.238262 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.238447 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.738401557 +0000 UTC m=+130.418093033 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.340020 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.340349 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.840333188 +0000 UTC m=+130.520024654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.365720 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" event={"ID":"73f65af5-9773-4165-8621-bf20d9eefd27","Type":"ContainerStarted","Data":"30622538cedd7ae851af41bc8f8bec1cc6835aec8fed39f1e4604b27f11e2ecf"} Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.365778 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" event={"ID":"73f65af5-9773-4165-8621-bf20d9eefd27","Type":"ContainerStarted","Data":"3b3af3d0ac374afcd3d7137a639ed704612a02a1807a90243477cc2a53dcc34a"} Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.365797 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6ctk"] Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.365978 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t7fhv"] Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.368405 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.370515 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.375049 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7fhv"] Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.375096 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5pdpj"] Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.375204 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.377247 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.383529 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5pdpj"] Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.383569 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.383583 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.383628 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.383637 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7xdjv"] Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.392429 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.401991 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-75ff9f647d-ptnqv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.402325 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7xdjv"] Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.402406 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.402035 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.419881 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-866fcbc849-n4k9x" podStartSLOduration=109.419866118 podStartE2EDuration="1m49.419866118s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:52.419239943 +0000 UTC m=+130.098931409" watchObservedRunningTime="2025-12-11 16:54:52.419866118 +0000 UTC m=+130.099557584" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.441054 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.441234 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-catalog-content\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.441270 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl4v4\" (UniqueName: \"kubernetes.io/projected/591b7ef7-10f4-43e0-9927-bb5ce699351c-kube-api-access-kl4v4\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.441381 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdxv9\" (UniqueName: \"kubernetes.io/projected/12530ca5-956e-43fd-9b42-40f50a32c8b8-kube-api-access-zdxv9\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.441439 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhh5z\" (UniqueName: \"kubernetes.io/projected/1751cd13-4bda-4f90-9998-306a003b25ab-kube-api-access-hhh5z\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.441611 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-utilities\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.441683 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-catalog-content\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.441834 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:52.941815997 +0000 UTC m=+130.621507463 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.446111 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-utilities\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.446300 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-utilities\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.446418 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-catalog-content\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.466210 5109 patch_prober.go:28] interesting pod/packageserver-7d4fc7d867-c2bmn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.466293 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" podUID="584167c4-b6c8-4abb-80cd-7aaed553a463" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.43:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.477204 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-67c89758df-8rnw5" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549422 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l75hz\" (UniqueName: \"kubernetes.io/projected/08b1e05d-7b30-4e51-83c3-90892198f91d-kube-api-access-l75hz\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549460 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-utilities\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549492 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-catalog-content\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549514 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549559 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-utilities\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549597 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-utilities\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549614 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-utilities\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549632 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-catalog-content\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549648 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-catalog-content\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549665 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kl4v4\" (UniqueName: \"kubernetes.io/projected/591b7ef7-10f4-43e0-9927-bb5ce699351c-kube-api-access-kl4v4\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549690 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zdxv9\" (UniqueName: \"kubernetes.io/projected/12530ca5-956e-43fd-9b42-40f50a32c8b8-kube-api-access-zdxv9\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549708 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-catalog-content\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.549731 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hhh5z\" (UniqueName: \"kubernetes.io/projected/1751cd13-4bda-4f90-9998-306a003b25ab-kube-api-access-hhh5z\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.550386 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-utilities\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.550693 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-catalog-content\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.550939 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.050927742 +0000 UTC m=+130.730619208 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.551296 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-utilities\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.551495 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-utilities\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.551771 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-catalog-content\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.552155 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-catalog-content\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.607599 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdxv9\" (UniqueName: \"kubernetes.io/projected/12530ca5-956e-43fd-9b42-40f50a32c8b8-kube-api-access-zdxv9\") pod \"community-operators-v6ctk\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.617266 5109 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-lth2t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 16:54:52 crc kubenswrapper[5109]: [-]has-synced failed: reason withheld Dec 11 16:54:52 crc kubenswrapper[5109]: [+]process-running ok Dec 11 16:54:52 crc kubenswrapper[5109]: healthz check failed Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.617325 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podUID="2e862e11-c81e-4697-bee2-8cb2b0ba590e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.643588 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhh5z\" (UniqueName: \"kubernetes.io/projected/1751cd13-4bda-4f90-9998-306a003b25ab-kube-api-access-hhh5z\") pod \"community-operators-5pdpj\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.655333 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.655542 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l75hz\" (UniqueName: \"kubernetes.io/projected/08b1e05d-7b30-4e51-83c3-90892198f91d-kube-api-access-l75hz\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.655957 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-utilities\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.656025 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-utilities\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.656073 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.156048299 +0000 UTC m=+130.835739765 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.656136 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl4v4\" (UniqueName: \"kubernetes.io/projected/591b7ef7-10f4-43e0-9927-bb5ce699351c-kube-api-access-kl4v4\") pod \"certified-operators-t7fhv\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.656150 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-catalog-content\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.656454 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-catalog-content\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.697990 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.701329 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.709497 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l75hz\" (UniqueName: \"kubernetes.io/projected/08b1e05d-7b30-4e51-83c3-90892198f91d-kube-api-access-l75hz\") pod \"certified-operators-7xdjv\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.721008 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.726536 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" event={"ID":"2d87a27d-ab14-48d9-a685-c43beae0f998","Type":"ContainerStarted","Data":"3417163f81fc76c36e14cabf6e5fb1e93765f44f666ed0cfca2fbee34c197879"} Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.737622 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.752186 5109 generic.go:358] "Generic (PLEG): container finished" podID="99387cea-c853-41ad-8926-39327379f170" containerID="c3a49be7c1286b87feb1651ae54c1a9acfc3aab9d04bc7012fedd61313a200a0" exitCode=0 Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.752336 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" event={"ID":"99387cea-c853-41ad-8926-39327379f170","Type":"ContainerDied","Data":"c3a49be7c1286b87feb1651ae54c1a9acfc3aab9d04bc7012fedd61313a200a0"} Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.758247 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.758555 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.258542333 +0000 UTC m=+130.938233799 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.785985 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" event={"ID":"68d5a689-abae-41b3-8c8b-0a9255fc276c","Type":"ContainerStarted","Data":"3f07f04d51703f951b6ddf8e21ff6bcfe0cba3fb892722a0954ee7fbe2be0177"} Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.794187 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" gracePeriod=30 Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.809928 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-7d4fc7d867-c2bmn" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.859505 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.861136 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.36112003 +0000 UTC m=+131.040811496 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.894217 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" podStartSLOduration=109.894198818 podStartE2EDuration="1m49.894198818s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:54:52.893665415 +0000 UTC m=+130.573356881" watchObservedRunningTime="2025-12-11 16:54:52.894198818 +0000 UTC m=+130.573890274" Dec 11 16:54:52 crc kubenswrapper[5109]: I1211 16:54:52.963024 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:52 crc kubenswrapper[5109]: E1211 16:54:52.963451 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.463433829 +0000 UTC m=+131.143125295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.072683 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.073443 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.573426744 +0000 UTC m=+131.253118210 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.115261 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55708: no serving certificate available for the kubelet" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.177695 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.178090 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.67807484 +0000 UTC m=+131.357766306 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.278521 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.279341 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.779324014 +0000 UTC m=+131.459015480 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.380320 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.380812 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.880799924 +0000 UTC m=+131.560491390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.385465 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v6ctk"] Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.483346 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.483626 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:53.983611075 +0000 UTC m=+131.663302541 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.584837 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.585224 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.085199098 +0000 UTC m=+131.764890564 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.609173 5109 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-lth2t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 16:54:53 crc kubenswrapper[5109]: [-]has-synced failed: reason withheld Dec 11 16:54:53 crc kubenswrapper[5109]: [+]process-running ok Dec 11 16:54:53 crc kubenswrapper[5109]: healthz check failed Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.609225 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podUID="2e862e11-c81e-4697-bee2-8cb2b0ba590e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.661256 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t7fhv"] Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.685792 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.686195 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.186175905 +0000 UTC m=+131.865867381 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.717636 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-txfkl"] Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.736991 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txfkl"] Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.737142 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.741181 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.769895 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7xdjv"] Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.788913 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.789357 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.289343176 +0000 UTC m=+131.969034642 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.792939 5109 patch_prober.go:28] interesting pod/oauth-openshift-66458b6674-mr25z container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.31:6443/healthz\": context deadline exceeded" start-of-body= Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.793008 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" podUID="693595ff-e600-48ce-9d44-2f996cc27307" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.31:6443/healthz\": context deadline exceeded" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.805020 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5pdpj"] Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.822249 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerStarted","Data":"b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85"} Dec 11 16:54:53 crc kubenswrapper[5109]: W1211 16:54:53.831752 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1751cd13_4bda_4f90_9998_306a003b25ab.slice/crio-fdb07aec719d711bdfc9cd03bab02d09c95c019aed94c616f93b26c870557b04 WatchSource:0}: Error finding container fdb07aec719d711bdfc9cd03bab02d09c95c019aed94c616f93b26c870557b04: Status 404 returned error can't find the container with id fdb07aec719d711bdfc9cd03bab02d09c95c019aed94c616f93b26c870557b04 Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.835953 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerStarted","Data":"7ccd45c7567db516dd1ed7083f3dbfb1f31f8e54fceecbbc4780edfd61289617"} Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.835986 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7fhv" event={"ID":"591b7ef7-10f4-43e0-9927-bb5ce699351c","Type":"ContainerStarted","Data":"f8eb8d7af2768f7390b51e5f4f4aefa9d0c08da6e3b1ae37a434dce748de2390"} Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.890469 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.890653 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.39062507 +0000 UTC m=+132.070316536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.890882 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-catalog-content\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.891090 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-utilities\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.891161 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.891227 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qp65\" (UniqueName: \"kubernetes.io/projected/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-kube-api-access-4qp65\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.891475 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.391467771 +0000 UTC m=+132.071159237 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.992922 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.993279 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-utilities\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.993393 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4qp65\" (UniqueName: \"kubernetes.io/projected/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-kube-api-access-4qp65\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.993613 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-catalog-content\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.994004 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-catalog-content\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:53 crc kubenswrapper[5109]: E1211 16:54:53.994068 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.494051887 +0000 UTC m=+132.173743353 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:53 crc kubenswrapper[5109]: I1211 16:54:53.996656 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-utilities\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.029054 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qp65\" (UniqueName: \"kubernetes.io/projected/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-kube-api-access-4qp65\") pod \"redhat-marketplace-txfkl\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.090440 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.102376 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.102757 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.60272303 +0000 UTC m=+132.282414496 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.105975 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pfq9w"] Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.132836 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.149350 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfq9w"] Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.160288 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.203266 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.203832 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjc6q\" (UniqueName: \"kubernetes.io/projected/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-kube-api-access-fjc6q\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.203952 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-utilities\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.204108 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-catalog-content\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.204306 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.704285292 +0000 UTC m=+132.383976758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.311618 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.311652 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-catalog-content\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.311713 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fjc6q\" (UniqueName: \"kubernetes.io/projected/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-kube-api-access-fjc6q\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.311763 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-utilities\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.312251 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.812239178 +0000 UTC m=+132.491930634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.312635 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-utilities\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.312705 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-catalog-content\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.343343 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjc6q\" (UniqueName: \"kubernetes.io/projected/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-kube-api-access-fjc6q\") pod \"redhat-marketplace-pfq9w\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.352398 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.413037 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.414086 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:54.914067906 +0000 UTC m=+132.593759372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.515191 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99387cea-c853-41ad-8926-39327379f170-secret-volume\") pod \"99387cea-c853-41ad-8926-39327379f170\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.515348 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f62vb\" (UniqueName: \"kubernetes.io/projected/99387cea-c853-41ad-8926-39327379f170-kube-api-access-f62vb\") pod \"99387cea-c853-41ad-8926-39327379f170\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.515577 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99387cea-c853-41ad-8926-39327379f170-config-volume\") pod \"99387cea-c853-41ad-8926-39327379f170\" (UID: \"99387cea-c853-41ad-8926-39327379f170\") " Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.515963 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.516307 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.016295594 +0000 UTC m=+132.695987050 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.518311 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/99387cea-c853-41ad-8926-39327379f170-config-volume" (OuterVolumeSpecName: "config-volume") pod "99387cea-c853-41ad-8926-39327379f170" (UID: "99387cea-c853-41ad-8926-39327379f170"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.525399 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99387cea-c853-41ad-8926-39327379f170-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "99387cea-c853-41ad-8926-39327379f170" (UID: "99387cea-c853-41ad-8926-39327379f170"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.535910 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99387cea-c853-41ad-8926-39327379f170-kube-api-access-f62vb" (OuterVolumeSpecName: "kube-api-access-f62vb") pod "99387cea-c853-41ad-8926-39327379f170" (UID: "99387cea-c853-41ad-8926-39327379f170"). InnerVolumeSpecName "kube-api-access-f62vb". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.536599 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txfkl"] Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.609335 5109 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-lth2t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 16:54:54 crc kubenswrapper[5109]: [-]has-synced failed: reason withheld Dec 11 16:54:54 crc kubenswrapper[5109]: [+]process-running ok Dec 11 16:54:54 crc kubenswrapper[5109]: healthz check failed Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.609402 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podUID="2e862e11-c81e-4697-bee2-8cb2b0ba590e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.613909 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.617181 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.617471 5109 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/99387cea-c853-41ad-8926-39327379f170-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.617487 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-f62vb\" (UniqueName: \"kubernetes.io/projected/99387cea-c853-41ad-8926-39327379f170-kube-api-access-f62vb\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.617495 5109 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/99387cea-c853-41ad-8926-39327379f170-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.617544 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.117531358 +0000 UTC m=+132.797222824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.689949 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.690731 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="99387cea-c853-41ad-8926-39327379f170" containerName="collect-profiles" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.690777 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="99387cea-c853-41ad-8926-39327379f170" containerName="collect-profiles" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.690885 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="99387cea-c853-41ad-8926-39327379f170" containerName="collect-profiles" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.694426 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.696651 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler\"/\"kube-root-ca.crt\"" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.699444 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.702950 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler\"/\"installer-sa-dockercfg-qpkss\"" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.719092 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.719645 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.219609512 +0000 UTC m=+132.899301058 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.820817 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.821090 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be691493-d391-44f3-8d8d-d63d438b8998-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.821151 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be691493-d391-44f3-8d8d-d63d438b8998-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.821260 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.321241506 +0000 UTC m=+133.000932972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.834259 5109 generic.go:358] "Generic (PLEG): container finished" podID="1751cd13-4bda-4f90-9998-306a003b25ab" containerID="f6c6f7cd694f658ee619f679e87a27ddc0482807988ce8aea450047ffe134d27" exitCode=0 Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.834352 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdpj" event={"ID":"1751cd13-4bda-4f90-9998-306a003b25ab","Type":"ContainerDied","Data":"f6c6f7cd694f658ee619f679e87a27ddc0482807988ce8aea450047ffe134d27"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.834378 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdpj" event={"ID":"1751cd13-4bda-4f90-9998-306a003b25ab","Type":"ContainerStarted","Data":"fdb07aec719d711bdfc9cd03bab02d09c95c019aed94c616f93b26c870557b04"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.835698 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" event={"ID":"99387cea-c853-41ad-8926-39327379f170","Type":"ContainerDied","Data":"1bb468f0191664e88328af6ba5be9ffc64e4262df5ff7236b4c3737526b0205d"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.835718 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bb468f0191664e88328af6ba5be9ffc64e4262df5ff7236b4c3737526b0205d" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.835714 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424525-4rxtv" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.841610 5109 generic.go:358] "Generic (PLEG): container finished" podID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerID="1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a" exitCode=0 Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.841797 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txfkl" event={"ID":"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363","Type":"ContainerDied","Data":"1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.841826 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txfkl" event={"ID":"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363","Type":"ContainerStarted","Data":"e104531d7ff56beae3ee5ffa0e3210ca03b83fc8b61dd3bbad955223b3f237f2"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.854903 5109 generic.go:358] "Generic (PLEG): container finished" podID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerID="b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85" exitCode=0 Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.855050 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerDied","Data":"b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.860403 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfq9w"] Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.861913 5109 generic.go:358] "Generic (PLEG): container finished" podID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerID="e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621" exitCode=0 Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.864388 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7fhv" event={"ID":"591b7ef7-10f4-43e0-9927-bb5ce699351c","Type":"ContainerDied","Data":"e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621"} Dec 11 16:54:54 crc kubenswrapper[5109]: W1211 16:54:54.868873 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod634b1a50_cd7a_4673_b3bd_d49ab2a1114e.slice/crio-6cb6e2f2b9495301a2c3f57b150f852269d71e7daa41b3728ff058ac9e511797 WatchSource:0}: Error finding container 6cb6e2f2b9495301a2c3f57b150f852269d71e7daa41b3728ff058ac9e511797: Status 404 returned error can't find the container with id 6cb6e2f2b9495301a2c3f57b150f852269d71e7daa41b3728ff058ac9e511797 Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.876200 5109 generic.go:358] "Generic (PLEG): container finished" podID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerID="c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0" exitCode=0 Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.877192 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xdjv" event={"ID":"08b1e05d-7b30-4e51-83c3-90892198f91d","Type":"ContainerDied","Data":"c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.877222 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xdjv" event={"ID":"08b1e05d-7b30-4e51-83c3-90892198f91d","Type":"ContainerStarted","Data":"50e7cd9cd1bf6352f7710ab27632b9f989108dff92a9bd9990a2cb4d9c6d5ef0"} Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.922559 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.922612 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be691493-d391-44f3-8d8d-d63d438b8998-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.922656 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be691493-d391-44f3-8d8d-d63d438b8998-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.923036 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be691493-d391-44f3-8d8d-d63d438b8998-kubelet-dir\") pod \"revision-pruner-6-crc\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:54 crc kubenswrapper[5109]: E1211 16:54:54.923203 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.423187366 +0000 UTC m=+133.102878832 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:54 crc kubenswrapper[5109]: I1211 16:54:54.973385 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be691493-d391-44f3-8d8d-d63d438b8998-kube-api-access\") pod \"revision-pruner-6-crc\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.011248 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.023628 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.024684 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.524655746 +0000 UTC m=+133.204347212 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.097625 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w9lmv"] Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.102765 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.104847 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.115177 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w9lmv"] Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.195349 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.195833 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.695812817 +0000 UTC m=+133.375504283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.297197 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.297490 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.797473482 +0000 UTC m=+133.477164948 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.297535 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfxwh\" (UniqueName: \"kubernetes.io/projected/d76f4925-c474-4764-9e50-1597f45a32b6-kube-api-access-jfxwh\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.297598 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-utilities\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.297659 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.297716 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-catalog-content\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.298069 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.798059126 +0000 UTC m=+133.477750592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.399091 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.399387 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.899356381 +0000 UTC m=+133.579047847 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.399614 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-utilities\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.399729 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.399863 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-catalog-content\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.399949 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jfxwh\" (UniqueName: \"kubernetes.io/projected/d76f4925-c474-4764-9e50-1597f45a32b6-kube-api-access-jfxwh\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.400382 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-utilities\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.400609 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-catalog-content\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.400850 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:55.900834567 +0000 UTC m=+133.580526023 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.418296 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfxwh\" (UniqueName: \"kubernetes.io/projected/d76f4925-c474-4764-9e50-1597f45a32b6-kube-api-access-jfxwh\") pod \"redhat-operators-w9lmv\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.496981 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nrqhg"] Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.501162 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.501433 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.001417994 +0000 UTC m=+133.681109460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.510246 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.523184 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrqhg"] Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.557195 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.593822 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler/revision-pruner-6-crc"] Dec 11 16:54:55 crc kubenswrapper[5109]: W1211 16:54:55.599651 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbe691493_d391_44f3_8d8d_d63d438b8998.slice/crio-4f1bd45731e24450b1113ef4fac9579e05de5cc24c653a45cbedef56f6070ae3 WatchSource:0}: Error finding container 4f1bd45731e24450b1113ef4fac9579e05de5cc24c653a45cbedef56f6070ae3: Status 404 returned error can't find the container with id 4f1bd45731e24450b1113ef4fac9579e05de5cc24c653a45cbedef56f6070ae3 Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.602065 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-catalog-content\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.602146 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-utilities\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.602407 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.602447 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n26jq\" (UniqueName: \"kubernetes.io/projected/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-kube-api-access-n26jq\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.602866 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.102853303 +0000 UTC m=+133.782544769 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.611393 5109 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-lth2t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 16:54:55 crc kubenswrapper[5109]: [-]has-synced failed: reason withheld Dec 11 16:54:55 crc kubenswrapper[5109]: [+]process-running ok Dec 11 16:54:55 crc kubenswrapper[5109]: healthz check failed Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.611483 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podUID="2e862e11-c81e-4697-bee2-8cb2b0ba590e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.703492 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.703808 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.20378127 +0000 UTC m=+133.883472736 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.704228 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.704267 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n26jq\" (UniqueName: \"kubernetes.io/projected/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-kube-api-access-n26jq\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.704318 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-catalog-content\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.705188 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.205168233 +0000 UTC m=+133.884859699 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.705876 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-utilities\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.706557 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-catalog-content\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.710179 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55718: no serving certificate available for the kubelet" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.712200 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-utilities\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.722709 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n26jq\" (UniqueName: \"kubernetes.io/projected/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-kube-api-access-n26jq\") pod \"redhat-operators-nrqhg\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.799697 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-5777786469-52jsw" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.807349 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.807487 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.307464683 +0000 UTC m=+133.987156149 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.808003 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.809408 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.309391979 +0000 UTC m=+133.989083515 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.826163 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.908632 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.908874 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.40885739 +0000 UTC m=+134.088548856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.909282 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:55 crc kubenswrapper[5109]: E1211 16:54:55.909558 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.409550817 +0000 UTC m=+134.089242283 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.916494 5109 generic.go:358] "Generic (PLEG): container finished" podID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerID="024ec6d69b690d4496e3b3f5cc484934437f249e9f7fc9bb1cfb2b438166833c" exitCode=0 Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.916726 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfq9w" event={"ID":"634b1a50-cd7a-4673-b3bd-d49ab2a1114e","Type":"ContainerDied","Data":"024ec6d69b690d4496e3b3f5cc484934437f249e9f7fc9bb1cfb2b438166833c"} Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.916764 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfq9w" event={"ID":"634b1a50-cd7a-4673-b3bd-d49ab2a1114e","Type":"ContainerStarted","Data":"6cb6e2f2b9495301a2c3f57b150f852269d71e7daa41b3728ff058ac9e511797"} Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.922459 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"be691493-d391-44f3-8d8d-d63d438b8998","Type":"ContainerStarted","Data":"4f1bd45731e24450b1113ef4fac9579e05de5cc24c653a45cbedef56f6070ae3"} Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.957960 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.958019 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.966166 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.981986 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.982031 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.995054 5109 patch_prober.go:28] interesting pod/console-64d44f6ddf-hvlqh container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 11 16:54:55 crc kubenswrapper[5109]: I1211 16:54:55.995152 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-console/console-64d44f6ddf-hvlqh" podUID="e4dc2bdf-6e49-4012-9f1c-d2af161dfd6d" containerName="console" probeResult="failure" output="Get \"https://10.217.0.21:8443/health\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.016022 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.016433 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.516419407 +0000 UTC m=+134.196110873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.018544 5109 patch_prober.go:28] interesting pod/downloads-747b44746d-bf7pr container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.018594 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-747b44746d-bf7pr" podUID="ce9e44ba-a2e2-443b-b265-aa312d55d7ca" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.029029 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w9lmv"] Dec 11 16:54:56 crc kubenswrapper[5109]: W1211 16:54:56.046089 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd76f4925_c474_4764_9e50_1597f45a32b6.slice/crio-a78cce0154ad46366334460dc0db4ca0c4dfb803dfe6ff1f22b178dfb6b877a4 WatchSource:0}: Error finding container a78cce0154ad46366334460dc0db4ca0c4dfb803dfe6ff1f22b178dfb6b877a4: Status 404 returned error can't find the container with id a78cce0154ad46366334460dc0db4ca0c4dfb803dfe6ff1f22b178dfb6b877a4 Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.094989 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nrqhg"] Dec 11 16:54:56 crc kubenswrapper[5109]: W1211 16:54:56.108721 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c4ce6e5_c194_4c26_a2c1_d61a0542af3d.slice/crio-9b0a768865f60fdbef198c4ad661f8ae9dac10331aa5c14173114dfde4cb4cc1 WatchSource:0}: Error finding container 9b0a768865f60fdbef198c4ad661f8ae9dac10331aa5c14173114dfde4cb4cc1: Status 404 returned error can't find the container with id 9b0a768865f60fdbef198c4ad661f8ae9dac10331aa5c14173114dfde4cb4cc1 Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.117754 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.118307 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.618295417 +0000 UTC m=+134.297986873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.219052 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.219264 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.719233112 +0000 UTC m=+134.398924608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.219426 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.219912 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.719904269 +0000 UTC m=+134.399595735 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.240255 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.240301 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.245643 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.266912 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-9ddfb9f55-47r72" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.326642 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.327112 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.827079229 +0000 UTC m=+134.506770705 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.428869 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.429354 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:56.929337928 +0000 UTC m=+134.609029404 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.529706 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.530018 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:57.029997267 +0000 UTC m=+134.709688723 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.609274 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.612293 5109 patch_prober.go:28] interesting pod/router-default-68cf44c8b8-lth2t container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 11 16:54:56 crc kubenswrapper[5109]: [-]has-synced failed: reason withheld Dec 11 16:54:56 crc kubenswrapper[5109]: [+]process-running ok Dec 11 16:54:56 crc kubenswrapper[5109]: healthz check failed Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.612390 5109 prober.go:120] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" podUID="2e862e11-c81e-4697-bee2-8cb2b0ba590e" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.631815 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.632233 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:57.132213705 +0000 UTC m=+134.811905171 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.733178 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.733590 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:57.23357442 +0000 UTC m=+134.913265886 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.768519 5109 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.834874 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.835192 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:57.335181183 +0000 UTC m=+135.014872649 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.938375 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.938898 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:57.438871367 +0000 UTC m=+135.118562833 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.939171 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:56 crc kubenswrapper[5109]: E1211 16:54:56.939534 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName: nodeName:}" failed. No retries permitted until 2025-12-11 16:54:57.439522814 +0000 UTC m=+135.119214280 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "image-registry-66587d64c8-nkxtk" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.951852 5109 generic.go:358] "Generic (PLEG): container finished" podID="d76f4925-c474-4764-9e50-1597f45a32b6" containerID="5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4" exitCode=0 Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.952016 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9lmv" event={"ID":"d76f4925-c474-4764-9e50-1597f45a32b6","Type":"ContainerDied","Data":"5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4"} Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.952043 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9lmv" event={"ID":"d76f4925-c474-4764-9e50-1597f45a32b6","Type":"ContainerStarted","Data":"a78cce0154ad46366334460dc0db4ca0c4dfb803dfe6ff1f22b178dfb6b877a4"} Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.958592 5109 generic.go:358] "Generic (PLEG): container finished" podID="be691493-d391-44f3-8d8d-d63d438b8998" containerID="c5c4f8f69045f4920b3eb189b66ca21a6bc8383b48b3abdbc1e861fe1fb8e1b3" exitCode=0 Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.958678 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"be691493-d391-44f3-8d8d-d63d438b8998","Type":"ContainerDied","Data":"c5c4f8f69045f4920b3eb189b66ca21a6bc8383b48b3abdbc1e861fe1fb8e1b3"} Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.961767 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" event={"ID":"68d5a689-abae-41b3-8c8b-0a9255fc276c","Type":"ContainerStarted","Data":"1a104dca1f005d472e761b2c04ffd0565fb00faec00be0b7d5b0c71225ded180"} Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.961800 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" event={"ID":"68d5a689-abae-41b3-8c8b-0a9255fc276c","Type":"ContainerStarted","Data":"107b5e199934d8d49eaa6713d919cf9298b2bb182e934d1c24a3630979119c9e"} Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.971675 5109 generic.go:358] "Generic (PLEG): container finished" podID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerID="92d03fccc279074df8d380ec9372f9dbf494c9dcde67692c039dd26e5ed56ee6" exitCode=0 Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.973387 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrqhg" event={"ID":"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d","Type":"ContainerDied","Data":"92d03fccc279074df8d380ec9372f9dbf494c9dcde67692c039dd26e5ed56ee6"} Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.973449 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrqhg" event={"ID":"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d","Type":"ContainerStarted","Data":"9b0a768865f60fdbef198c4ad661f8ae9dac10331aa5c14173114dfde4cb4cc1"} Dec 11 16:54:56 crc kubenswrapper[5109]: I1211 16:54:56.983134 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-8596bd845d-gb4zv" Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.040573 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:57 crc kubenswrapper[5109]: E1211 16:54:57.042164 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2 podName:9e9b5059-1b3e-4067-a63d-2952cbe863af nodeName:}" failed. No retries permitted until 2025-12-11 16:54:57.542142471 +0000 UTC m=+135.221833947 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.095833 5109 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-11T16:54:56.768550919Z","UUID":"7c2d2bf9-e95f-4349-b7f0-0e1dadc18335","Handler":null,"Name":"","Endpoint":""} Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.114179 5109 csi_plugin.go:106] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.114215 5109 csi_plugin.go:119] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.145487 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.178060 5109 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.178097 5109 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b1264ac67579ad07e7e9003054d44fe40dd55285a4b2f7dc74e48be1aee0868a/globalmount\"" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.248348 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-66587d64c8-nkxtk\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.348914 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"9e9b5059-1b3e-4067-a63d-2952cbe863af\" (UID: \"9e9b5059-1b3e-4067-a63d-2952cbe863af\") " Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.357719 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (OuterVolumeSpecName: "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2") pod "9e9b5059-1b3e-4067-a63d-2952cbe863af" (UID: "9e9b5059-1b3e-4067-a63d-2952cbe863af"). InnerVolumeSpecName "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2". PluginName "kubernetes.io/csi", VolumeGIDValue "" Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.429051 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:54:57 crc kubenswrapper[5109]: I1211 16:54:57.593374 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Dec 11 16:54:59 crc kubenswrapper[5109]: I1211 16:54:59.987832 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Dec 11 16:54:59 crc kubenswrapper[5109]: I1211 16:54:59.988199 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-747b44746d-bf7pr" Dec 11 16:54:59 crc kubenswrapper[5109]: I1211 16:54:59.988229 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-nkxtk"] Dec 11 16:54:59 crc kubenswrapper[5109]: I1211 16:54:59.988267 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-77chb" Dec 11 16:54:59 crc kubenswrapper[5109]: I1211 16:54:59.988364 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.006604 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.007569 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.032022 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e9b5059-1b3e-4067-a63d-2952cbe863af" path="/var/lib/kubelet/pods/9e9b5059-1b3e-4067-a63d-2952cbe863af/volumes" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.037951 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.037994 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" event={"ID":"eead6019-41c4-4ac6-9226-8ec4c456db13","Type":"ContainerStarted","Data":"63ce238e0168bb3f1a6aa70488fb3da556a43d96b9d0bf1b56faddf8c18b07eb"} Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.046331 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" event={"ID":"68d5a689-abae-41b3-8c8b-0a9255fc276c","Type":"ContainerStarted","Data":"dbef0fe14fd34d3425db58f6c290c65550c95867bc13d6f1d0ae6a5577ff02f9"} Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.058255 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-68cf44c8b8-lth2t" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.109147 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df5bf83-da14-48ab-8cc5-5c9918418025-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.112514 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df5bf83-da14-48ab-8cc5-5c9918418025-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.216172 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df5bf83-da14-48ab-8cc5-5c9918418025-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.216677 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df5bf83-da14-48ab-8cc5-5c9918418025-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.217131 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df5bf83-da14-48ab-8cc5-5c9918418025-kubelet-dir\") pod \"revision-pruner-11-crc\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.253053 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df5bf83-da14-48ab-8cc5-5c9918418025-kube-api-access\") pod \"revision-pruner-11-crc\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.332225 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.366790 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.418110 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be691493-d391-44f3-8d8d-d63d438b8998-kubelet-dir\") pod \"be691493-d391-44f3-8d8d-d63d438b8998\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.418217 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be691493-d391-44f3-8d8d-d63d438b8998-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "be691493-d391-44f3-8d8d-d63d438b8998" (UID: "be691493-d391-44f3-8d8d-d63d438b8998"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.418538 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be691493-d391-44f3-8d8d-d63d438b8998-kube-api-access\") pod \"be691493-d391-44f3-8d8d-d63d438b8998\" (UID: \"be691493-d391-44f3-8d8d-d63d438b8998\") " Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.419003 5109 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/be691493-d391-44f3-8d8d-d63d438b8998-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.428058 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be691493-d391-44f3-8d8d-d63d438b8998-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "be691493-d391-44f3-8d8d-d63d438b8998" (UID: "be691493-d391-44f3-8d8d-d63d438b8998"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:00 crc kubenswrapper[5109]: E1211 16:55:00.507133 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:00 crc kubenswrapper[5109]: E1211 16:55:00.508833 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:00 crc kubenswrapper[5109]: E1211 16:55:00.510101 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:00 crc kubenswrapper[5109]: E1211 16:55:00.510134 5109 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.516995 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-11-crc"] Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.520360 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/be691493-d391-44f3-8d8d-d63d438b8998-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:00 crc kubenswrapper[5109]: W1211 16:55:00.525784 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4df5bf83_da14_48ab_8cc5_5c9918418025.slice/crio-96ffb387481b22997b063bd4e8c00b6556c1f0731878e571413eb4d7f443d00d WatchSource:0}: Error finding container 96ffb387481b22997b063bd4e8c00b6556c1f0731878e571413eb4d7f443d00d: Status 404 returned error can't find the container with id 96ffb387481b22997b063bd4e8c00b6556c1f0731878e571413eb4d7f443d00d Dec 11 16:55:00 crc kubenswrapper[5109]: I1211 16:55:00.859920 5109 ???:1] "http: TLS handshake error from 192.168.126.11:47692: no serving certificate available for the kubelet" Dec 11 16:55:01 crc kubenswrapper[5109]: I1211 16:55:01.054784 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"4df5bf83-da14-48ab-8cc5-5c9918418025","Type":"ContainerStarted","Data":"96ffb387481b22997b063bd4e8c00b6556c1f0731878e571413eb4d7f443d00d"} Dec 11 16:55:01 crc kubenswrapper[5109]: I1211 16:55:01.057456 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/revision-pruner-6-crc" Dec 11 16:55:01 crc kubenswrapper[5109]: I1211 16:55:01.057455 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/revision-pruner-6-crc" event={"ID":"be691493-d391-44f3-8d8d-d63d438b8998","Type":"ContainerDied","Data":"4f1bd45731e24450b1113ef4fac9579e05de5cc24c653a45cbedef56f6070ae3"} Dec 11 16:55:01 crc kubenswrapper[5109]: I1211 16:55:01.057626 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f1bd45731e24450b1113ef4fac9579e05de5cc24c653a45cbedef56f6070ae3" Dec 11 16:55:01 crc kubenswrapper[5109]: I1211 16:55:01.884818 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-pp75t" podStartSLOduration=19.88479231 podStartE2EDuration="19.88479231s" podCreationTimestamp="2025-12-11 16:54:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:01.881240292 +0000 UTC m=+139.560931818" watchObservedRunningTime="2025-12-11 16:55:01.88479231 +0000 UTC m=+139.564483816" Dec 11 16:55:02 crc kubenswrapper[5109]: I1211 16:55:02.067071 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" event={"ID":"eead6019-41c4-4ac6-9226-8ec4c456db13","Type":"ContainerStarted","Data":"ddfcef0be238b0f543e5629758bb06f24ae93d271be9a1f78fadaba269b1b9de"} Dec 11 16:55:02 crc kubenswrapper[5109]: I1211 16:55:02.067977 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:55:02 crc kubenswrapper[5109]: I1211 16:55:02.090185 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" podStartSLOduration=119.090172688 podStartE2EDuration="1m59.090172688s" podCreationTimestamp="2025-12-11 16:53:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:02.086670372 +0000 UTC m=+139.766361838" watchObservedRunningTime="2025-12-11 16:55:02.090172688 +0000 UTC m=+139.769864154" Dec 11 16:55:03 crc kubenswrapper[5109]: I1211 16:55:03.076267 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"4df5bf83-da14-48ab-8cc5-5c9918418025","Type":"ContainerStarted","Data":"c509f4d4fb2455712fb0269ed99f25d068dc2fda90cd2dcf4bf8586c76282f85"} Dec 11 16:55:04 crc kubenswrapper[5109]: I1211 16:55:04.101727 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-11-crc" podStartSLOduration=7.101710095 podStartE2EDuration="7.101710095s" podCreationTimestamp="2025-12-11 16:54:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:04.099310325 +0000 UTC m=+141.779001801" watchObservedRunningTime="2025-12-11 16:55:04.101710095 +0000 UTC m=+141.781401601" Dec 11 16:55:05 crc kubenswrapper[5109]: I1211 16:55:05.094333 5109 generic.go:358] "Generic (PLEG): container finished" podID="4df5bf83-da14-48ab-8cc5-5c9918418025" containerID="c509f4d4fb2455712fb0269ed99f25d068dc2fda90cd2dcf4bf8586c76282f85" exitCode=0 Dec 11 16:55:05 crc kubenswrapper[5109]: I1211 16:55:05.094510 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"4df5bf83-da14-48ab-8cc5-5c9918418025","Type":"ContainerDied","Data":"c509f4d4fb2455712fb0269ed99f25d068dc2fda90cd2dcf4bf8586c76282f85"} Dec 11 16:55:05 crc kubenswrapper[5109]: I1211 16:55:05.982840 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:55:05 crc kubenswrapper[5109]: I1211 16:55:05.990099 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-64d44f6ddf-hvlqh" Dec 11 16:55:06 crc kubenswrapper[5109]: I1211 16:55:06.546865 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 16:55:09 crc kubenswrapper[5109]: I1211 16:55:09.832464 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-9zkbs"] Dec 11 16:55:09 crc kubenswrapper[5109]: I1211 16:55:09.833378 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" podUID="c7efc5a6-365a-411a-8cc9-575086c1c169" containerName="controller-manager" containerID="cri-o://4b6fbed6af02d089962b5aebe82e539ad168a4c742fc2ff980cdbf0e3be91aef" gracePeriod=30 Dec 11 16:55:09 crc kubenswrapper[5109]: I1211 16:55:09.852239 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk"] Dec 11 16:55:09 crc kubenswrapper[5109]: I1211 16:55:09.852998 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" podUID="e52494b7-da4c-4aa1-82cc-285cc203d719" containerName="route-controller-manager" containerID="cri-o://e752a9ba4d3d1b106f525cb419736f25a555d5005f334a1ba848fa3db90f3d1b" gracePeriod=30 Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.080366 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.126394 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-11-crc" event={"ID":"4df5bf83-da14-48ab-8cc5-5c9918418025","Type":"ContainerDied","Data":"96ffb387481b22997b063bd4e8c00b6556c1f0731878e571413eb4d7f443d00d"} Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.126434 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96ffb387481b22997b063bd4e8c00b6556c1f0731878e571413eb4d7f443d00d" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.126507 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-11-crc" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.128280 5109 generic.go:358] "Generic (PLEG): container finished" podID="c7efc5a6-365a-411a-8cc9-575086c1c169" containerID="4b6fbed6af02d089962b5aebe82e539ad168a4c742fc2ff980cdbf0e3be91aef" exitCode=0 Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.128460 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" event={"ID":"c7efc5a6-365a-411a-8cc9-575086c1c169","Type":"ContainerDied","Data":"4b6fbed6af02d089962b5aebe82e539ad168a4c742fc2ff980cdbf0e3be91aef"} Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.130339 5109 generic.go:358] "Generic (PLEG): container finished" podID="e52494b7-da4c-4aa1-82cc-285cc203d719" containerID="e752a9ba4d3d1b106f525cb419736f25a555d5005f334a1ba848fa3db90f3d1b" exitCode=0 Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.130423 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" event={"ID":"e52494b7-da4c-4aa1-82cc-285cc203d719","Type":"ContainerDied","Data":"e752a9ba4d3d1b106f525cb419736f25a555d5005f334a1ba848fa3db90f3d1b"} Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.171950 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df5bf83-da14-48ab-8cc5-5c9918418025-kube-api-access\") pod \"4df5bf83-da14-48ab-8cc5-5c9918418025\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.173387 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df5bf83-da14-48ab-8cc5-5c9918418025-kubelet-dir\") pod \"4df5bf83-da14-48ab-8cc5-5c9918418025\" (UID: \"4df5bf83-da14-48ab-8cc5-5c9918418025\") " Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.173492 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4df5bf83-da14-48ab-8cc5-5c9918418025-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4df5bf83-da14-48ab-8cc5-5c9918418025" (UID: "4df5bf83-da14-48ab-8cc5-5c9918418025"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.174370 5109 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4df5bf83-da14-48ab-8cc5-5c9918418025-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.180485 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4df5bf83-da14-48ab-8cc5-5c9918418025-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4df5bf83-da14-48ab-8cc5-5c9918418025" (UID: "4df5bf83-da14-48ab-8cc5-5c9918418025"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.276605 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4df5bf83-da14-48ab-8cc5-5c9918418025-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:10 crc kubenswrapper[5109]: E1211 16:55:10.498899 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:10 crc kubenswrapper[5109]: E1211 16:55:10.501515 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:10 crc kubenswrapper[5109]: E1211 16:55:10.506808 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:10 crc kubenswrapper[5109]: E1211 16:55:10.506891 5109 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.683118 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.683189 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.683427 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.683453 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.684694 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-nginx-conf\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.689479 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwt8b\" (UniqueName: \"kubernetes.io/projected/17b87002-b798-480a-8e17-83053d698239-kube-api-access-gwt8b\") pod \"network-check-target-fhkjl\" (UID: \"17b87002-b798-480a-8e17-83053d698239\") " pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.689533 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7w75\" (UniqueName: \"kubernetes.io/projected/f863fff9-286a-45fa-b8f0-8a86994b8440-kube-api-access-l7w75\") pod \"network-check-source-5bb8f5cd97-xdvz5\" (UID: \"f863fff9-286a-45fa-b8f0-8a86994b8440\") " pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.689940 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/6a9ae5f6-97bd-46ac-bafa-ca1b4452a141-networking-console-plugin-cert\") pod \"networking-console-plugin-5ff7774fd9-nljh6\" (UID: \"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141\") " pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.795386 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.817053 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" Dec 11 16:55:10 crc kubenswrapper[5109]: I1211 16:55:10.822790 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" Dec 11 16:55:11 crc kubenswrapper[5109]: I1211 16:55:11.089035 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:55:11 crc kubenswrapper[5109]: I1211 16:55:11.092938 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ad4ed271-a1a8-408d-b043-99cf044fb573-metrics-certs\") pod \"network-metrics-daemon-fp2cw\" (UID: \"ad4ed271-a1a8-408d-b043-99cf044fb573\") " pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:55:11 crc kubenswrapper[5109]: I1211 16:55:11.103708 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-fp2cw" Dec 11 16:55:11 crc kubenswrapper[5109]: I1211 16:55:11.119889 5109 ???:1] "http: TLS handshake error from 192.168.126.11:47654: no serving certificate available for the kubelet" Dec 11 16:55:12 crc kubenswrapper[5109]: I1211 16:55:12.367452 5109 patch_prober.go:28] interesting pod/route-controller-manager-776cdc94d6-gp4qk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 11 16:55:12 crc kubenswrapper[5109]: I1211 16:55:12.367578 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" podUID="e52494b7-da4c-4aa1-82cc-285cc203d719" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 11 16:55:15 crc kubenswrapper[5109]: I1211 16:55:15.563170 5109 patch_prober.go:28] interesting pod/controller-manager-65b6cccf98-9zkbs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 11 16:55:15 crc kubenswrapper[5109]: I1211 16:55:15.563890 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" podUID="c7efc5a6-365a-411a-8cc9-575086c1c169" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 11 16:55:20 crc kubenswrapper[5109]: E1211 16:55:20.498700 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:20 crc kubenswrapper[5109]: E1211 16:55:20.501872 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:20 crc kubenswrapper[5109]: E1211 16:55:20.503823 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" cmd=["/bin/bash","-c","test -f /ready/ready"] Dec 11 16:55:20 crc kubenswrapper[5109]: E1211 16:55:20.503903 5109 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerName="kube-multus-additional-cni-plugins" probeResult="unknown" Dec 11 16:55:22 crc kubenswrapper[5109]: I1211 16:55:22.367640 5109 patch_prober.go:28] interesting pod/route-controller-manager-776cdc94d6-gp4qk container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Dec 11 16:55:22 crc kubenswrapper[5109]: I1211 16:55:22.367706 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" podUID="e52494b7-da4c-4aa1-82cc-285cc203d719" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Dec 11 16:55:23 crc kubenswrapper[5109]: I1211 16:55:23.831317 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-77f986bd66-7kbk8" Dec 11 16:55:24 crc kubenswrapper[5109]: I1211 16:55:24.088210 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 16:55:24 crc kubenswrapper[5109]: I1211 16:55:24.232652 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-nbqqf_4eed92a0-6f24-424d-9be1-1f7e336ce8a9/kube-multus-additional-cni-plugins/0.log" Dec 11 16:55:24 crc kubenswrapper[5109]: I1211 16:55:24.232697 5109 generic.go:358] "Generic (PLEG): container finished" podID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" exitCode=137 Dec 11 16:55:24 crc kubenswrapper[5109]: I1211 16:55:24.232802 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" event={"ID":"4eed92a0-6f24-424d-9be1-1f7e336ce8a9","Type":"ContainerDied","Data":"b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122"} Dec 11 16:55:25 crc kubenswrapper[5109]: I1211 16:55:25.562332 5109 patch_prober.go:28] interesting pod/controller-manager-65b6cccf98-9zkbs container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Dec 11 16:55:25 crc kubenswrapper[5109]: I1211 16:55:25.562721 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" podUID="c7efc5a6-365a-411a-8cc9-575086c1c169" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.825809 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-nbqqf_4eed92a0-6f24-424d-9be1-1f7e336ce8a9/kube-multus-additional-cni-plugins/0.log" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.826975 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.842711 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.868344 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.878405 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6d556dc944-9cq4k"] Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879257 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c7efc5a6-365a-411a-8cc9-575086c1c169" containerName="controller-manager" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879274 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7efc5a6-365a-411a-8cc9-575086c1c169" containerName="controller-manager" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879288 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerName="kube-multus-additional-cni-plugins" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879295 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerName="kube-multus-additional-cni-plugins" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879318 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="4df5bf83-da14-48ab-8cc5-5c9918418025" containerName="pruner" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879325 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="4df5bf83-da14-48ab-8cc5-5c9918418025" containerName="pruner" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879342 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="be691493-d391-44f3-8d8d-d63d438b8998" containerName="pruner" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879349 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="be691493-d391-44f3-8d8d-d63d438b8998" containerName="pruner" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879368 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e52494b7-da4c-4aa1-82cc-285cc203d719" containerName="route-controller-manager" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879375 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="e52494b7-da4c-4aa1-82cc-285cc203d719" containerName="route-controller-manager" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879473 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" containerName="kube-multus-additional-cni-plugins" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879489 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="c7efc5a6-365a-411a-8cc9-575086c1c169" containerName="controller-manager" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879499 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="be691493-d391-44f3-8d8d-d63d438b8998" containerName="pruner" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879508 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="4df5bf83-da14-48ab-8cc5-5c9918418025" containerName="pruner" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.879516 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="e52494b7-da4c-4aa1-82cc-285cc203d719" containerName="route-controller-manager" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923251 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-client-ca\") pod \"c7efc5a6-365a-411a-8cc9-575086c1c169\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923290 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7efc5a6-365a-411a-8cc9-575086c1c169-tmp\") pod \"c7efc5a6-365a-411a-8cc9-575086c1c169\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923319 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-client-ca\") pod \"e52494b7-da4c-4aa1-82cc-285cc203d719\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923344 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-ready\") pod \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923409 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-config\") pod \"e52494b7-da4c-4aa1-82cc-285cc203d719\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923531 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7efc5a6-365a-411a-8cc9-575086c1c169-serving-cert\") pod \"c7efc5a6-365a-411a-8cc9-575086c1c169\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923601 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q9qt\" (UniqueName: \"kubernetes.io/projected/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-kube-api-access-6q9qt\") pod \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923638 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-tuning-conf-dir\") pod \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923659 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5zvn\" (UniqueName: \"kubernetes.io/projected/e52494b7-da4c-4aa1-82cc-285cc203d719-kube-api-access-v5zvn\") pod \"e52494b7-da4c-4aa1-82cc-285cc203d719\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923704 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psk95\" (UniqueName: \"kubernetes.io/projected/c7efc5a6-365a-411a-8cc9-575086c1c169-kube-api-access-psk95\") pod \"c7efc5a6-365a-411a-8cc9-575086c1c169\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923758 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-config\") pod \"c7efc5a6-365a-411a-8cc9-575086c1c169\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923776 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e52494b7-da4c-4aa1-82cc-285cc203d719-tmp\") pod \"e52494b7-da4c-4aa1-82cc-285cc203d719\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923795 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-proxy-ca-bundles\") pod \"c7efc5a6-365a-411a-8cc9-575086c1c169\" (UID: \"c7efc5a6-365a-411a-8cc9-575086c1c169\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923814 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52494b7-da4c-4aa1-82cc-285cc203d719-serving-cert\") pod \"e52494b7-da4c-4aa1-82cc-285cc203d719\" (UID: \"e52494b7-da4c-4aa1-82cc-285cc203d719\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.923835 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist\") pod \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\" (UID: \"4eed92a0-6f24-424d-9be1-1f7e336ce8a9\") " Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.924038 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-client-ca" (OuterVolumeSpecName: "client-ca") pod "e52494b7-da4c-4aa1-82cc-285cc203d719" (UID: "e52494b7-da4c-4aa1-82cc-285cc203d719"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.924080 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-config" (OuterVolumeSpecName: "config") pod "e52494b7-da4c-4aa1-82cc-285cc203d719" (UID: "e52494b7-da4c-4aa1-82cc-285cc203d719"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.924177 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.924194 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e52494b7-da4c-4aa1-82cc-285cc203d719-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.924777 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "4eed92a0-6f24-424d-9be1-1f7e336ce8a9" (UID: "4eed92a0-6f24-424d-9be1-1f7e336ce8a9"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.925049 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e52494b7-da4c-4aa1-82cc-285cc203d719-tmp" (OuterVolumeSpecName: "tmp") pod "e52494b7-da4c-4aa1-82cc-285cc203d719" (UID: "e52494b7-da4c-4aa1-82cc-285cc203d719"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.925489 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-ready" (OuterVolumeSpecName: "ready") pod "4eed92a0-6f24-424d-9be1-1f7e336ce8a9" (UID: "4eed92a0-6f24-424d-9be1-1f7e336ce8a9"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.925521 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "4eed92a0-6f24-424d-9be1-1f7e336ce8a9" (UID: "4eed92a0-6f24-424d-9be1-1f7e336ce8a9"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.925848 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7efc5a6-365a-411a-8cc9-575086c1c169-tmp" (OuterVolumeSpecName: "tmp") pod "c7efc5a6-365a-411a-8cc9-575086c1c169" (UID: "c7efc5a6-365a-411a-8cc9-575086c1c169"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.927578 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-config" (OuterVolumeSpecName: "config") pod "c7efc5a6-365a-411a-8cc9-575086c1c169" (UID: "c7efc5a6-365a-411a-8cc9-575086c1c169"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.928637 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-client-ca" (OuterVolumeSpecName: "client-ca") pod "c7efc5a6-365a-411a-8cc9-575086c1c169" (UID: "c7efc5a6-365a-411a-8cc9-575086c1c169"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.929755 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c7efc5a6-365a-411a-8cc9-575086c1c169" (UID: "c7efc5a6-365a-411a-8cc9-575086c1c169"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.930047 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-kube-api-access-6q9qt" (OuterVolumeSpecName: "kube-api-access-6q9qt") pod "4eed92a0-6f24-424d-9be1-1f7e336ce8a9" (UID: "4eed92a0-6f24-424d-9be1-1f7e336ce8a9"). InnerVolumeSpecName "kube-api-access-6q9qt". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.933105 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e52494b7-da4c-4aa1-82cc-285cc203d719-kube-api-access-v5zvn" (OuterVolumeSpecName: "kube-api-access-v5zvn") pod "e52494b7-da4c-4aa1-82cc-285cc203d719" (UID: "e52494b7-da4c-4aa1-82cc-285cc203d719"). InnerVolumeSpecName "kube-api-access-v5zvn". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.933282 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7efc5a6-365a-411a-8cc9-575086c1c169-kube-api-access-psk95" (OuterVolumeSpecName: "kube-api-access-psk95") pod "c7efc5a6-365a-411a-8cc9-575086c1c169" (UID: "c7efc5a6-365a-411a-8cc9-575086c1c169"). InnerVolumeSpecName "kube-api-access-psk95". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.936128 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e52494b7-da4c-4aa1-82cc-285cc203d719-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e52494b7-da4c-4aa1-82cc-285cc203d719" (UID: "e52494b7-da4c-4aa1-82cc-285cc203d719"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.936332 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7efc5a6-365a-411a-8cc9-575086c1c169-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c7efc5a6-365a-411a-8cc9-575086c1c169" (UID: "c7efc5a6-365a-411a-8cc9-575086c1c169"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.939691 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d556dc944-9cq4k"] Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.941292 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:26 crc kubenswrapper[5109]: I1211 16:55:26.978931 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.009869 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.010030 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026005 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-proxy-ca-bundles\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026066 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pt4c9\" (UniqueName: \"kubernetes.io/projected/d4a48a49-3fc3-4096-aad3-e98160f224fa-kube-api-access-pt4c9\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026102 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-config\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026122 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4a48a49-3fc3-4096-aad3-e98160f224fa-tmp\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026137 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-client-ca\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026151 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a48a49-3fc3-4096-aad3-e98160f224fa-serving-cert\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026223 5109 reconciler_common.go:299] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026234 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026243 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c7efc5a6-365a-411a-8cc9-575086c1c169-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026251 5109 reconciler_common.go:299] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-ready\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026259 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7efc5a6-365a-411a-8cc9-575086c1c169-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026267 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6q9qt\" (UniqueName: \"kubernetes.io/projected/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-kube-api-access-6q9qt\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026319 5109 reconciler_common.go:299] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4eed92a0-6f24-424d-9be1-1f7e336ce8a9-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026327 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-v5zvn\" (UniqueName: \"kubernetes.io/projected/e52494b7-da4c-4aa1-82cc-285cc203d719-kube-api-access-v5zvn\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026336 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-psk95\" (UniqueName: \"kubernetes.io/projected/c7efc5a6-365a-411a-8cc9-575086c1c169-kube-api-access-psk95\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026344 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026352 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/e52494b7-da4c-4aa1-82cc-285cc203d719-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026359 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c7efc5a6-365a-411a-8cc9-575086c1c169-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.026367 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e52494b7-da4c-4aa1-82cc-285cc203d719-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.127344 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-client-ca\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.127426 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-proxy-ca-bundles\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.127479 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4082104-5770-4f9c-a78b-6116f05589a6-tmp\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.128576 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-proxy-ca-bundles\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.128691 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-pt4c9\" (UniqueName: \"kubernetes.io/projected/d4a48a49-3fc3-4096-aad3-e98160f224fa-kube-api-access-pt4c9\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.128732 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr97m\" (UniqueName: \"kubernetes.io/projected/d4082104-5770-4f9c-a78b-6116f05589a6-kube-api-access-zr97m\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.128779 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-config\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.128799 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-config\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.134065 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4a48a49-3fc3-4096-aad3-e98160f224fa-tmp\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.134095 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-client-ca\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.130560 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-config\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.134115 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a48a49-3fc3-4096-aad3-e98160f224fa-serving-cert\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.134225 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4082104-5770-4f9c-a78b-6116f05589a6-serving-cert\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.134759 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4a48a49-3fc3-4096-aad3-e98160f224fa-tmp\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.135309 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-client-ca\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.140654 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a48a49-3fc3-4096-aad3-e98160f224fa-serving-cert\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.147758 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-pt4c9\" (UniqueName: \"kubernetes.io/projected/d4a48a49-3fc3-4096-aad3-e98160f224fa-kube-api-access-pt4c9\") pod \"controller-manager-6d556dc944-9cq4k\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.242083 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-client-ca\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.242133 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4082104-5770-4f9c-a78b-6116f05589a6-tmp\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.242179 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zr97m\" (UniqueName: \"kubernetes.io/projected/d4082104-5770-4f9c-a78b-6116f05589a6-kube-api-access-zr97m\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.242204 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-config\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.242229 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4082104-5770-4f9c-a78b-6116f05589a6-serving-cert\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.242869 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4082104-5770-4f9c-a78b-6116f05589a6-tmp\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.243598 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-client-ca\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.243911 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-config\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.247436 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4082104-5770-4f9c-a78b-6116f05589a6-serving-cert\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.263516 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr97m\" (UniqueName: \"kubernetes.io/projected/d4082104-5770-4f9c-a78b-6116f05589a6-kube-api-access-zr97m\") pod \"route-controller-manager-7bdd58d8d6-648f8\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.266910 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" event={"ID":"c7efc5a6-365a-411a-8cc9-575086c1c169","Type":"ContainerDied","Data":"c05d6e7f66922abdd01b49890d6eb4bc292f5f78eb7058bbba82e7f642c42794"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.266967 5109 scope.go:117] "RemoveContainer" containerID="4b6fbed6af02d089962b5aebe82e539ad168a4c742fc2ff980cdbf0e3be91aef" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.267142 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-65b6cccf98-9zkbs" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.270376 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-nbqqf_4eed92a0-6f24-424d-9be1-1f7e336ce8a9/kube-multus-additional-cni-plugins/0.log" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.270648 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" event={"ID":"4eed92a0-6f24-424d-9be1-1f7e336ce8a9","Type":"ContainerDied","Data":"c51d5399e4312d80df519f40cafa445d5f6ff8237b195392a2c76bab68f34e41"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.270708 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-nbqqf" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.284513 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-9zkbs"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.286222 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-65b6cccf98-9zkbs"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.294189 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerStarted","Data":"0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.298952 5109 scope.go:117] "RemoveContainer" containerID="b2f7c5e493c47258af47c4181d6091eb742d26de774fa858597e2eb7e05f4122" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.302238 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-nbqqf"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.304683 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-nbqqf"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.315492 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" event={"ID":"e52494b7-da4c-4aa1-82cc-285cc203d719","Type":"ContainerDied","Data":"37f77bbbddf8c5c87a17471e7206047fa6d2661a70ea89379cbab0f4b171b08b"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.315608 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.318535 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrqhg" event={"ID":"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d","Type":"ContainerStarted","Data":"e63fdec881e3ce29accee172651a4f84892cb8c49c06932ec600cbbf85dc4821"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.325950 5109 scope.go:117] "RemoveContainer" containerID="e752a9ba4d3d1b106f525cb419736f25a555d5005f334a1ba848fa3db90f3d1b" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.335625 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9lmv" event={"ID":"d76f4925-c474-4764-9e50-1597f45a32b6","Type":"ContainerStarted","Data":"a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.346299 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.349400 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-fp2cw"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.364623 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7fhv" event={"ID":"591b7ef7-10f4-43e0-9927-bb5ce699351c","Type":"ContainerStarted","Data":"b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.373605 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdpj" event={"ID":"1751cd13-4bda-4f90-9998-306a003b25ab","Type":"ContainerStarted","Data":"993e1fb3073fafbdf1263a62bc088c890ac0af75bd20b18da26e41f166b856bc"} Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.382948 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.385286 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776cdc94d6-gp4qk"] Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.432925 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:27 crc kubenswrapper[5109]: W1211 16:55:27.448999 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podad4ed271_a1a8_408d_b043_99cf044fb573.slice/crio-d59b75a97b1c3a1029c10b5b7a182fbcbb835cea5f9290f9c710b3e4e71133f4 WatchSource:0}: Error finding container d59b75a97b1c3a1029c10b5b7a182fbcbb835cea5f9290f9c710b3e4e71133f4: Status 404 returned error can't find the container with id d59b75a97b1c3a1029c10b5b7a182fbcbb835cea5f9290f9c710b3e4e71133f4 Dec 11 16:55:27 crc kubenswrapper[5109]: W1211 16:55:27.550167 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6a9ae5f6_97bd_46ac_bafa_ca1b4452a141.slice/crio-6eb944871bb4b4806ec52b210bb3fce2387b1a09eba6749cb9be3eb6745b836b WatchSource:0}: Error finding container 6eb944871bb4b4806ec52b210bb3fce2387b1a09eba6749cb9be3eb6745b836b: Status 404 returned error can't find the container with id 6eb944871bb4b4806ec52b210bb3fce2387b1a09eba6749cb9be3eb6745b836b Dec 11 16:55:27 crc kubenswrapper[5109]: W1211 16:55:27.576491 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17b87002_b798_480a_8e17_83053d698239.slice/crio-787e1446017da18646b77d544c42edfd192334d9450e3158db416119496d8172 WatchSource:0}: Error finding container 787e1446017da18646b77d544c42edfd192334d9450e3158db416119496d8172: Status 404 returned error can't find the container with id 787e1446017da18646b77d544c42edfd192334d9450e3158db416119496d8172 Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.661857 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6d556dc944-9cq4k"] Dec 11 16:55:27 crc kubenswrapper[5109]: W1211 16:55:27.697547 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd4a48a49_3fc3_4096_aad3_e98160f224fa.slice/crio-09bd8450f7af1bf72a291d002c3c781cd343ab3e18485468415c936b80c438a6 WatchSource:0}: Error finding container 09bd8450f7af1bf72a291d002c3c781cd343ab3e18485468415c936b80c438a6: Status 404 returned error can't find the container with id 09bd8450f7af1bf72a291d002c3c781cd343ab3e18485468415c936b80c438a6 Dec 11 16:55:27 crc kubenswrapper[5109]: I1211 16:55:27.712773 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8"] Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.387997 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" event={"ID":"f863fff9-286a-45fa-b8f0-8a86994b8440","Type":"ContainerStarted","Data":"18dd7b4fe742d749e2dfaff27aa38c8ec487e81b020fbbfd11e2b96e3ee0b736"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.388903 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-5bb8f5cd97-xdvz5" event={"ID":"f863fff9-286a-45fa-b8f0-8a86994b8440","Type":"ContainerStarted","Data":"c401be072057cef01e46d6580a26c1d2cf3136392acd6a3ec8505b67e02328bc"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.391554 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" event={"ID":"d4a48a49-3fc3-4096-aad3-e98160f224fa","Type":"ContainerStarted","Data":"331b9dda8c325bcc996a81ed4afa31c6b11a468abbdec3d105bacb15b9ca3ed3"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.391600 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" event={"ID":"d4a48a49-3fc3-4096-aad3-e98160f224fa","Type":"ContainerStarted","Data":"09bd8450f7af1bf72a291d002c3c781cd343ab3e18485468415c936b80c438a6"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.392560 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.398603 5109 generic.go:358] "Generic (PLEG): container finished" podID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerID="0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.398703 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerDied","Data":"0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.412671 5109 generic.go:358] "Generic (PLEG): container finished" podID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerID="7db66df61f0fb5f9d15de57bab0bfc3c6b1a5330ee9019cb4cee469a50f5d6b0" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.412905 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfq9w" event={"ID":"634b1a50-cd7a-4673-b3bd-d49ab2a1114e","Type":"ContainerDied","Data":"7db66df61f0fb5f9d15de57bab0bfc3c6b1a5330ee9019cb4cee469a50f5d6b0"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.418916 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" event={"ID":"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141","Type":"ContainerStarted","Data":"5f278c820dd2f639bbeb61743c99def703a3c374793b9cf2600605cca86e59cf"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.418968 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-5ff7774fd9-nljh6" event={"ID":"6a9ae5f6-97bd-46ac-bafa-ca1b4452a141","Type":"ContainerStarted","Data":"6eb944871bb4b4806ec52b210bb3fce2387b1a09eba6749cb9be3eb6745b836b"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.441904 5109 generic.go:358] "Generic (PLEG): container finished" podID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerID="e63fdec881e3ce29accee172651a4f84892cb8c49c06932ec600cbbf85dc4821" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.442146 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrqhg" event={"ID":"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d","Type":"ContainerDied","Data":"e63fdec881e3ce29accee172651a4f84892cb8c49c06932ec600cbbf85dc4821"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.445828 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" podStartSLOduration=19.445806424 podStartE2EDuration="19.445806424s" podCreationTimestamp="2025-12-11 16:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:28.44083468 +0000 UTC m=+166.120526156" watchObservedRunningTime="2025-12-11 16:55:28.445806424 +0000 UTC m=+166.125497890" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.451448 5109 generic.go:358] "Generic (PLEG): container finished" podID="d76f4925-c474-4764-9e50-1597f45a32b6" containerID="a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.451533 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9lmv" event={"ID":"d76f4925-c474-4764-9e50-1597f45a32b6","Type":"ContainerDied","Data":"a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.456074 5109 generic.go:358] "Generic (PLEG): container finished" podID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerID="b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.456157 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7fhv" event={"ID":"591b7ef7-10f4-43e0-9927-bb5ce699351c","Type":"ContainerDied","Data":"b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.468240 5109 generic.go:358] "Generic (PLEG): container finished" podID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerID="60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.468460 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xdjv" event={"ID":"08b1e05d-7b30-4e51-83c3-90892198f91d","Type":"ContainerDied","Data":"60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.473819 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" event={"ID":"d4082104-5770-4f9c-a78b-6116f05589a6","Type":"ContainerStarted","Data":"969d360fd6cd8ca20593a400223272459fa8f104698661472d3126514632bf03"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.473859 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" event={"ID":"d4082104-5770-4f9c-a78b-6116f05589a6","Type":"ContainerStarted","Data":"b9694ae01f3d2cbacf7b33e2cbc53875f71757a4f7af5f2d812ac9fea5a3a020"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.475524 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" event={"ID":"ad4ed271-a1a8-408d-b043-99cf044fb573","Type":"ContainerStarted","Data":"09d270a8d80a656cf6123f9e6b64d566880de1fa2f119be5c19d15c3e59fc4e8"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.475550 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" event={"ID":"ad4ed271-a1a8-408d-b043-99cf044fb573","Type":"ContainerStarted","Data":"d59b75a97b1c3a1029c10b5b7a182fbcbb835cea5f9290f9c710b3e4e71133f4"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.479408 5109 generic.go:358] "Generic (PLEG): container finished" podID="1751cd13-4bda-4f90-9998-306a003b25ab" containerID="993e1fb3073fafbdf1263a62bc088c890ac0af75bd20b18da26e41f166b856bc" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.479492 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdpj" event={"ID":"1751cd13-4bda-4f90-9998-306a003b25ab","Type":"ContainerDied","Data":"993e1fb3073fafbdf1263a62bc088c890ac0af75bd20b18da26e41f166b856bc"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.482673 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" event={"ID":"17b87002-b798-480a-8e17-83053d698239","Type":"ContainerStarted","Data":"87a6bd08022a5c093cce80abe2cfac48e007818b0187ceb41628bad5d83fc612"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.482701 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-fhkjl" event={"ID":"17b87002-b798-480a-8e17-83053d698239","Type":"ContainerStarted","Data":"787e1446017da18646b77d544c42edfd192334d9450e3158db416119496d8172"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.484523 5109 generic.go:358] "Generic (PLEG): container finished" podID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerID="7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5" exitCode=0 Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.484648 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txfkl" event={"ID":"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363","Type":"ContainerDied","Data":"7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5"} Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.517821 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.527910 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.565229 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" podStartSLOduration=19.565208277 podStartE2EDuration="19.565208277s" podCreationTimestamp="2025-12-11 16:55:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:28.562610333 +0000 UTC m=+166.242301819" watchObservedRunningTime="2025-12-11 16:55:28.565208277 +0000 UTC m=+166.244899753" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.840909 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.875328 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.912643 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4eed92a0-6f24-424d-9be1-1f7e336ce8a9" path="/var/lib/kubelet/pods/4eed92a0-6f24-424d-9be1-1f7e336ce8a9/volumes" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.913500 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7efc5a6-365a-411a-8cc9-575086c1c169" path="/var/lib/kubelet/pods/c7efc5a6-365a-411a-8cc9-575086c1c169/volumes" Dec 11 16:55:28 crc kubenswrapper[5109]: I1211 16:55:28.914049 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e52494b7-da4c-4aa1-82cc-285cc203d719" path="/var/lib/kubelet/pods/e52494b7-da4c-4aa1-82cc-285cc203d719/volumes" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.012985 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d556dc944-9cq4k"] Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.018930 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8"] Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.493232 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7fhv" event={"ID":"591b7ef7-10f4-43e0-9927-bb5ce699351c","Type":"ContainerStarted","Data":"f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.495661 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xdjv" event={"ID":"08b1e05d-7b30-4e51-83c3-90892198f91d","Type":"ContainerStarted","Data":"a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.497334 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-fp2cw" event={"ID":"ad4ed271-a1a8-408d-b043-99cf044fb573","Type":"ContainerStarted","Data":"2471d1f3eeaac663cd5b30367dec16a6ead222ecae7bf3144f79cd0125f2ea02"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.499687 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdpj" event={"ID":"1751cd13-4bda-4f90-9998-306a003b25ab","Type":"ContainerStarted","Data":"7ab02dd030d9048ebe6fca705655d45e4b4fae25dbcc7d1120e76c1716a53c07"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.501467 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txfkl" event={"ID":"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363","Type":"ContainerStarted","Data":"947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.503563 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerStarted","Data":"92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.505642 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfq9w" event={"ID":"634b1a50-cd7a-4673-b3bd-d49ab2a1114e","Type":"ContainerStarted","Data":"a7c8e710f96d5b14dec9182f53e41a00dff8a342d023fab570334dcb8dbcbc85"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.507974 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrqhg" event={"ID":"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d","Type":"ContainerStarted","Data":"56272c96d2cd188288af93f0f82c691ba3c48eb076b1c089bbea32cc0a13ef22"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.510035 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9lmv" event={"ID":"d76f4925-c474-4764-9e50-1597f45a32b6","Type":"ContainerStarted","Data":"d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3"} Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.512705 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t7fhv" podStartSLOduration=6.615957064 podStartE2EDuration="38.512691069s" podCreationTimestamp="2025-12-11 16:54:51 +0000 UTC" firstStartedPulling="2025-12-11 16:54:54.864246283 +0000 UTC m=+132.543937749" lastFinishedPulling="2025-12-11 16:55:26.760980278 +0000 UTC m=+164.440671754" observedRunningTime="2025-12-11 16:55:29.510046193 +0000 UTC m=+167.189737659" watchObservedRunningTime="2025-12-11 16:55:29.512691069 +0000 UTC m=+167.192382535" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.529933 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5pdpj" podStartSLOduration=5.50341029 podStartE2EDuration="37.529912506s" podCreationTimestamp="2025-12-11 16:54:52 +0000 UTC" firstStartedPulling="2025-12-11 16:54:54.835353236 +0000 UTC m=+132.515044702" lastFinishedPulling="2025-12-11 16:55:26.861855442 +0000 UTC m=+164.541546918" observedRunningTime="2025-12-11 16:55:29.528394619 +0000 UTC m=+167.208086095" watchObservedRunningTime="2025-12-11 16:55:29.529912506 +0000 UTC m=+167.209603972" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.585433 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v6ctk" podStartSLOduration=6.658482655 podStartE2EDuration="38.585416424s" podCreationTimestamp="2025-12-11 16:54:51 +0000 UTC" firstStartedPulling="2025-12-11 16:54:54.85583394 +0000 UTC m=+132.535525406" lastFinishedPulling="2025-12-11 16:55:26.782767709 +0000 UTC m=+164.462459175" observedRunningTime="2025-12-11 16:55:29.583910497 +0000 UTC m=+167.263601963" watchObservedRunningTime="2025-12-11 16:55:29.585416424 +0000 UTC m=+167.265107890" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.585849 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nrqhg" podStartSLOduration=4.7289276099999995 podStartE2EDuration="34.585841645s" podCreationTimestamp="2025-12-11 16:54:55 +0000 UTC" firstStartedPulling="2025-12-11 16:54:56.973909167 +0000 UTC m=+134.653600633" lastFinishedPulling="2025-12-11 16:55:26.830823202 +0000 UTC m=+164.510514668" observedRunningTime="2025-12-11 16:55:29.550525778 +0000 UTC m=+167.230217244" watchObservedRunningTime="2025-12-11 16:55:29.585841645 +0000 UTC m=+167.265533111" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.608254 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7xdjv" podStartSLOduration=5.625301297 podStartE2EDuration="37.608236231s" podCreationTimestamp="2025-12-11 16:54:52 +0000 UTC" firstStartedPulling="2025-12-11 16:54:54.877044642 +0000 UTC m=+132.556736098" lastFinishedPulling="2025-12-11 16:55:26.859979546 +0000 UTC m=+164.539671032" observedRunningTime="2025-12-11 16:55:29.606711933 +0000 UTC m=+167.286403399" watchObservedRunningTime="2025-12-11 16:55:29.608236231 +0000 UTC m=+167.287927707" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.631289 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pfq9w" podStartSLOduration=4.788551462 podStartE2EDuration="35.631273072s" podCreationTimestamp="2025-12-11 16:54:54 +0000 UTC" firstStartedPulling="2025-12-11 16:54:55.917684434 +0000 UTC m=+133.597375900" lastFinishedPulling="2025-12-11 16:55:26.760406044 +0000 UTC m=+164.440097510" observedRunningTime="2025-12-11 16:55:29.630241577 +0000 UTC m=+167.309933063" watchObservedRunningTime="2025-12-11 16:55:29.631273072 +0000 UTC m=+167.310964538" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.647758 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-txfkl" podStartSLOduration=4.661682234 podStartE2EDuration="36.647727621s" podCreationTimestamp="2025-12-11 16:54:53 +0000 UTC" firstStartedPulling="2025-12-11 16:54:54.842621872 +0000 UTC m=+132.522313338" lastFinishedPulling="2025-12-11 16:55:26.828667249 +0000 UTC m=+164.508358725" observedRunningTime="2025-12-11 16:55:29.645980718 +0000 UTC m=+167.325672184" watchObservedRunningTime="2025-12-11 16:55:29.647727621 +0000 UTC m=+167.327419087" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.663404 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-fp2cw" podStartSLOduration=147.663387689 podStartE2EDuration="2m27.663387689s" podCreationTimestamp="2025-12-11 16:53:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:29.661521303 +0000 UTC m=+167.341212769" watchObservedRunningTime="2025-12-11 16:55:29.663387689 +0000 UTC m=+167.343079155" Dec 11 16:55:29 crc kubenswrapper[5109]: I1211 16:55:29.681932 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w9lmv" podStartSLOduration=4.703875678 podStartE2EDuration="34.68191362s" podCreationTimestamp="2025-12-11 16:54:55 +0000 UTC" firstStartedPulling="2025-12-11 16:54:56.952868475 +0000 UTC m=+134.632559931" lastFinishedPulling="2025-12-11 16:55:26.930906407 +0000 UTC m=+164.610597873" observedRunningTime="2025-12-11 16:55:29.679238104 +0000 UTC m=+167.358929570" watchObservedRunningTime="2025-12-11 16:55:29.68191362 +0000 UTC m=+167.361605076" Dec 11 16:55:30 crc kubenswrapper[5109]: I1211 16:55:30.515281 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" podUID="d4082104-5770-4f9c-a78b-6116f05589a6" containerName="route-controller-manager" containerID="cri-o://969d360fd6cd8ca20593a400223272459fa8f104698661472d3126514632bf03" gracePeriod=30 Dec 11 16:55:30 crc kubenswrapper[5109]: I1211 16:55:30.517300 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" podUID="d4a48a49-3fc3-4096-aad3-e98160f224fa" containerName="controller-manager" containerID="cri-o://331b9dda8c325bcc996a81ed4afa31c6b11a468abbdec3d105bacb15b9ca3ed3" gracePeriod=30 Dec 11 16:55:31 crc kubenswrapper[5109]: I1211 16:55:31.525593 5109 generic.go:358] "Generic (PLEG): container finished" podID="d4082104-5770-4f9c-a78b-6116f05589a6" containerID="969d360fd6cd8ca20593a400223272459fa8f104698661472d3126514632bf03" exitCode=0 Dec 11 16:55:31 crc kubenswrapper[5109]: I1211 16:55:31.525677 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" event={"ID":"d4082104-5770-4f9c-a78b-6116f05589a6","Type":"ContainerDied","Data":"969d360fd6cd8ca20593a400223272459fa8f104698661472d3126514632bf03"} Dec 11 16:55:31 crc kubenswrapper[5109]: I1211 16:55:31.527470 5109 generic.go:358] "Generic (PLEG): container finished" podID="d4a48a49-3fc3-4096-aad3-e98160f224fa" containerID="331b9dda8c325bcc996a81ed4afa31c6b11a468abbdec3d105bacb15b9ca3ed3" exitCode=0 Dec 11 16:55:31 crc kubenswrapper[5109]: I1211 16:55:31.527548 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" event={"ID":"d4a48a49-3fc3-4096-aad3-e98160f224fa","Type":"ContainerDied","Data":"331b9dda8c325bcc996a81ed4afa31c6b11a468abbdec3d105bacb15b9ca3ed3"} Dec 11 16:55:31 crc kubenswrapper[5109]: I1211 16:55:31.627677 5109 ???:1] "http: TLS handshake error from 192.168.126.11:51334: no serving certificate available for the kubelet" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.045920 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.050431 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.072515 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.073106 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d4082104-5770-4f9c-a78b-6116f05589a6" containerName="route-controller-manager" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.073127 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4082104-5770-4f9c-a78b-6116f05589a6" containerName="route-controller-manager" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.073138 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d4a48a49-3fc3-4096-aad3-e98160f224fa" containerName="controller-manager" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.073145 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a48a49-3fc3-4096-aad3-e98160f224fa" containerName="controller-manager" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.073243 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="d4082104-5770-4f9c-a78b-6116f05589a6" containerName="route-controller-manager" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.073258 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="d4a48a49-3fc3-4096-aad3-e98160f224fa" containerName="controller-manager" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.084337 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.096256 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5dd796b77f-l9zgm"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.100141 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.107027 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.124400 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4082104-5770-4f9c-a78b-6116f05589a6-tmp\") pod \"d4082104-5770-4f9c-a78b-6116f05589a6\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.124710 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-client-ca\") pod \"d4082104-5770-4f9c-a78b-6116f05589a6\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.124785 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-config\") pod \"d4a48a49-3fc3-4096-aad3-e98160f224fa\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.124817 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pt4c9\" (UniqueName: \"kubernetes.io/projected/d4a48a49-3fc3-4096-aad3-e98160f224fa-kube-api-access-pt4c9\") pod \"d4a48a49-3fc3-4096-aad3-e98160f224fa\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.124863 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-config\") pod \"d4082104-5770-4f9c-a78b-6116f05589a6\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.124922 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a48a49-3fc3-4096-aad3-e98160f224fa-serving-cert\") pod \"d4a48a49-3fc3-4096-aad3-e98160f224fa\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.124988 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4a48a49-3fc3-4096-aad3-e98160f224fa-tmp\") pod \"d4a48a49-3fc3-4096-aad3-e98160f224fa\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.125069 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-client-ca\") pod \"d4a48a49-3fc3-4096-aad3-e98160f224fa\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.125113 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-proxy-ca-bundles\") pod \"d4a48a49-3fc3-4096-aad3-e98160f224fa\" (UID: \"d4a48a49-3fc3-4096-aad3-e98160f224fa\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.125133 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zr97m\" (UniqueName: \"kubernetes.io/projected/d4082104-5770-4f9c-a78b-6116f05589a6-kube-api-access-zr97m\") pod \"d4082104-5770-4f9c-a78b-6116f05589a6\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.125153 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4082104-5770-4f9c-a78b-6116f05589a6-serving-cert\") pod \"d4082104-5770-4f9c-a78b-6116f05589a6\" (UID: \"d4082104-5770-4f9c-a78b-6116f05589a6\") " Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.127342 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4a48a49-3fc3-4096-aad3-e98160f224fa-tmp" (OuterVolumeSpecName: "tmp") pod "d4a48a49-3fc3-4096-aad3-e98160f224fa" (UID: "d4a48a49-3fc3-4096-aad3-e98160f224fa"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.127484 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "d4082104-5770-4f9c-a78b-6116f05589a6" (UID: "d4082104-5770-4f9c-a78b-6116f05589a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.127643 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4082104-5770-4f9c-a78b-6116f05589a6-tmp" (OuterVolumeSpecName: "tmp") pod "d4082104-5770-4f9c-a78b-6116f05589a6" (UID: "d4082104-5770-4f9c-a78b-6116f05589a6"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.127716 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-config" (OuterVolumeSpecName: "config") pod "d4082104-5770-4f9c-a78b-6116f05589a6" (UID: "d4082104-5770-4f9c-a78b-6116f05589a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.127860 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-client-ca" (OuterVolumeSpecName: "client-ca") pod "d4a48a49-3fc3-4096-aad3-e98160f224fa" (UID: "d4a48a49-3fc3-4096-aad3-e98160f224fa"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.127974 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "d4a48a49-3fc3-4096-aad3-e98160f224fa" (UID: "d4a48a49-3fc3-4096-aad3-e98160f224fa"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.128020 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-config" (OuterVolumeSpecName: "config") pod "d4a48a49-3fc3-4096-aad3-e98160f224fa" (UID: "d4a48a49-3fc3-4096-aad3-e98160f224fa"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.137977 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a48a49-3fc3-4096-aad3-e98160f224fa-kube-api-access-pt4c9" (OuterVolumeSpecName: "kube-api-access-pt4c9") pod "d4a48a49-3fc3-4096-aad3-e98160f224fa" (UID: "d4a48a49-3fc3-4096-aad3-e98160f224fa"). InnerVolumeSpecName "kube-api-access-pt4c9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.138196 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dd796b77f-l9zgm"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.145957 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4082104-5770-4f9c-a78b-6116f05589a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d4082104-5770-4f9c-a78b-6116f05589a6" (UID: "d4082104-5770-4f9c-a78b-6116f05589a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.150084 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4082104-5770-4f9c-a78b-6116f05589a6-kube-api-access-zr97m" (OuterVolumeSpecName: "kube-api-access-zr97m") pod "d4082104-5770-4f9c-a78b-6116f05589a6" (UID: "d4082104-5770-4f9c-a78b-6116f05589a6"). InnerVolumeSpecName "kube-api-access-zr97m". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.152255 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4a48a49-3fc3-4096-aad3-e98160f224fa-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d4a48a49-3fc3-4096-aad3-e98160f224fa" (UID: "d4a48a49-3fc3-4096-aad3-e98160f224fa"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227045 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvsx8\" (UniqueName: \"kubernetes.io/projected/b27c4651-d447-4fd1-8f8d-fc81e9778230-kube-api-access-rvsx8\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227096 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-serving-cert\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227116 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm85n\" (UniqueName: \"kubernetes.io/projected/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-kube-api-access-lm85n\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227157 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-client-ca\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227238 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-proxy-ca-bundles\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227308 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-config\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227335 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-tmp\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227353 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-client-ca\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227584 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b27c4651-d447-4fd1-8f8d-fc81e9778230-tmp\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227676 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-config\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227752 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b27c4651-d447-4fd1-8f8d-fc81e9778230-serving-cert\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227911 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227930 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zr97m\" (UniqueName: \"kubernetes.io/projected/d4082104-5770-4f9c-a78b-6116f05589a6-kube-api-access-zr97m\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227942 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4082104-5770-4f9c-a78b-6116f05589a6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227951 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4082104-5770-4f9c-a78b-6116f05589a6-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227974 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227984 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.227991 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-pt4c9\" (UniqueName: \"kubernetes.io/projected/d4a48a49-3fc3-4096-aad3-e98160f224fa-kube-api-access-pt4c9\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.228000 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d4082104-5770-4f9c-a78b-6116f05589a6-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.228008 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d4a48a49-3fc3-4096-aad3-e98160f224fa-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.228016 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d4a48a49-3fc3-4096-aad3-e98160f224fa-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.228024 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d4a48a49-3fc3-4096-aad3-e98160f224fa-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329020 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b27c4651-d447-4fd1-8f8d-fc81e9778230-tmp\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329098 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-config\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329137 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b27c4651-d447-4fd1-8f8d-fc81e9778230-serving-cert\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329171 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rvsx8\" (UniqueName: \"kubernetes.io/projected/b27c4651-d447-4fd1-8f8d-fc81e9778230-kube-api-access-rvsx8\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329192 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-serving-cert\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329214 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lm85n\" (UniqueName: \"kubernetes.io/projected/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-kube-api-access-lm85n\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329271 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-client-ca\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329298 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-proxy-ca-bundles\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329333 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-config\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329348 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-tmp\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.329366 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-client-ca\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.330215 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-client-ca\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.330463 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b27c4651-d447-4fd1-8f8d-fc81e9778230-tmp\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.331432 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-config\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.332364 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-client-ca\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.332496 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-tmp\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.333071 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-config\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.333677 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-proxy-ca-bundles\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.336478 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b27c4651-d447-4fd1-8f8d-fc81e9778230-serving-cert\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.336966 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-serving-cert\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.347636 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm85n\" (UniqueName: \"kubernetes.io/projected/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-kube-api-access-lm85n\") pod \"controller-manager-5dd796b77f-l9zgm\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.349059 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvsx8\" (UniqueName: \"kubernetes.io/projected/b27c4651-d447-4fd1-8f8d-fc81e9778230-kube-api-access-rvsx8\") pod \"route-controller-manager-8997cb6f7-6rt4d\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.407031 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.417223 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.562385 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" event={"ID":"d4082104-5770-4f9c-a78b-6116f05589a6","Type":"ContainerDied","Data":"b9694ae01f3d2cbacf7b33e2cbc53875f71757a4f7af5f2d812ac9fea5a3a020"} Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.562453 5109 scope.go:117] "RemoveContainer" containerID="969d360fd6cd8ca20593a400223272459fa8f104698661472d3126514632bf03" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.562656 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.568379 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" event={"ID":"d4a48a49-3fc3-4096-aad3-e98160f224fa","Type":"ContainerDied","Data":"09bd8450f7af1bf72a291d002c3c781cd343ab3e18485468415c936b80c438a6"} Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.568439 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6d556dc944-9cq4k" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.594399 5109 scope.go:117] "RemoveContainer" containerID="331b9dda8c325bcc996a81ed4afa31c6b11a468abbdec3d105bacb15b9ca3ed3" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.604140 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.606020 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7bdd58d8d6-648f8"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.612142 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5dd796b77f-l9zgm"] Dec 11 16:55:32 crc kubenswrapper[5109]: W1211 16:55:32.629103 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod06ae51de_2665_44c0_8a42_bfc7bb42e7d8.slice/crio-1e00f37db6a65a8ad705e78dc17a1b2e863b5a4b0b47a1fcc09df103c59a7239 WatchSource:0}: Error finding container 1e00f37db6a65a8ad705e78dc17a1b2e863b5a4b0b47a1fcc09df103c59a7239: Status 404 returned error can't find the container with id 1e00f37db6a65a8ad705e78dc17a1b2e863b5a4b0b47a1fcc09df103c59a7239 Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.630249 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6d556dc944-9cq4k"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.634674 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6d556dc944-9cq4k"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.663089 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d"] Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.699556 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.699604 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.702820 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.703010 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.722773 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.722824 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.738428 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.738466 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.915171 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4082104-5770-4f9c-a78b-6116f05589a6" path="/var/lib/kubelet/pods/d4082104-5770-4f9c-a78b-6116f05589a6/volumes" Dec 11 16:55:32 crc kubenswrapper[5109]: I1211 16:55:32.916086 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a48a49-3fc3-4096-aad3-e98160f224fa" path="/var/lib/kubelet/pods/d4a48a49-3fc3-4096-aad3-e98160f224fa/volumes" Dec 11 16:55:33 crc kubenswrapper[5109]: I1211 16:55:33.575775 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" event={"ID":"06ae51de-2665-44c0-8a42-bfc7bb42e7d8","Type":"ContainerStarted","Data":"1e00f37db6a65a8ad705e78dc17a1b2e863b5a4b0b47a1fcc09df103c59a7239"} Dec 11 16:55:33 crc kubenswrapper[5109]: I1211 16:55:33.578077 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" event={"ID":"b27c4651-d447-4fd1-8f8d-fc81e9778230","Type":"ContainerStarted","Data":"48b7e7ef3661efa32fdd3b9f1ae0d65f97ae5ae2f1c95abd1ac21d02ecad02a0"} Dec 11 16:55:33 crc kubenswrapper[5109]: I1211 16:55:33.796321 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Dec 11 16:55:33 crc kubenswrapper[5109]: I1211 16:55:33.993872 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Dec 11 16:55:33 crc kubenswrapper[5109]: I1211 16:55:33.994039 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:33 crc kubenswrapper[5109]: I1211 16:55:33.996310 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\"" Dec 11 16:55:33 crc kubenswrapper[5109]: I1211 16:55:33.996534 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-bqqnb\"" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.038540 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.042703 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.043866 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.054343 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.091570 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.091946 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.137720 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.153402 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.154239 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.255221 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.255614 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.255688 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kubelet-dir\") pod \"revision-pruner-12-crc\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.279583 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kube-api-access\") pod \"revision-pruner-12-crc\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.314219 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.614721 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.614992 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.665228 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.665313 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.666245 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:55:34 crc kubenswrapper[5109]: I1211 16:55:34.675517 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.411515 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.412407 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.558988 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.559534 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.593831 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" event={"ID":"b27c4651-d447-4fd1-8f8d-fc81e9778230","Type":"ContainerStarted","Data":"90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d"} Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.595894 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" event={"ID":"06ae51de-2665-44c0-8a42-bfc7bb42e7d8","Type":"ContainerStarted","Data":"f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23"} Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.596151 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-12-crc"] Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.604674 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:55:35 crc kubenswrapper[5109]: W1211 16:55:35.614079 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda7ef01bd_dacd_4a7a_a8a4_cc5222bc5596.slice/crio-0c29a23f300c3d13a154582ddb468322eac8be42d61ac24e221d9dd34d6b852a WatchSource:0}: Error finding container 0c29a23f300c3d13a154582ddb468322eac8be42d61ac24e221d9dd34d6b852a: Status 404 returned error can't find the container with id 0c29a23f300c3d13a154582ddb468322eac8be42d61ac24e221d9dd34d6b852a Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.646865 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.657203 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.827181 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.827542 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:55:35 crc kubenswrapper[5109]: I1211 16:55:35.870221 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.604041 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596","Type":"ContainerStarted","Data":"0c29a23f300c3d13a154582ddb468322eac8be42d61ac24e221d9dd34d6b852a"} Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.606242 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.606277 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.611407 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.611463 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.622538 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" podStartSLOduration=7.622519891 podStartE2EDuration="7.622519891s" podCreationTimestamp="2025-12-11 16:55:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:36.617624179 +0000 UTC m=+174.297315665" watchObservedRunningTime="2025-12-11 16:55:36.622519891 +0000 UTC m=+174.302211357" Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.639174 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" podStartSLOduration=7.639157384 podStartE2EDuration="7.639157384s" podCreationTimestamp="2025-12-11 16:55:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:36.636789255 +0000 UTC m=+174.316480721" watchObservedRunningTime="2025-12-11 16:55:36.639157384 +0000 UTC m=+174.318848850" Dec 11 16:55:36 crc kubenswrapper[5109]: I1211 16:55:36.650354 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:55:37 crc kubenswrapper[5109]: I1211 16:55:37.538524 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7xdjv"] Dec 11 16:55:37 crc kubenswrapper[5109]: I1211 16:55:37.538801 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7xdjv" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="registry-server" containerID="cri-o://a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47" gracePeriod=2 Dec 11 16:55:37 crc kubenswrapper[5109]: I1211 16:55:37.612806 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596","Type":"ContainerStarted","Data":"b022ba5e9b0dd216bbf92cb8a69cac404c550c185d7710f0e161868cdb5ef67c"} Dec 11 16:55:37 crc kubenswrapper[5109]: I1211 16:55:37.632430 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-12-crc" podStartSLOduration=4.632410511 podStartE2EDuration="4.632410511s" podCreationTimestamp="2025-12-11 16:55:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:37.631209221 +0000 UTC m=+175.310900727" watchObservedRunningTime="2025-12-11 16:55:37.632410511 +0000 UTC m=+175.312101987" Dec 11 16:55:37 crc kubenswrapper[5109]: I1211 16:55:37.736852 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5pdpj"] Dec 11 16:55:37 crc kubenswrapper[5109]: I1211 16:55:37.737195 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5pdpj" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="registry-server" containerID="cri-o://7ab02dd030d9048ebe6fca705655d45e4b4fae25dbcc7d1120e76c1716a53c07" gracePeriod=2 Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.496773 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.627063 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l75hz\" (UniqueName: \"kubernetes.io/projected/08b1e05d-7b30-4e51-83c3-90892198f91d-kube-api-access-l75hz\") pod \"08b1e05d-7b30-4e51-83c3-90892198f91d\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.627114 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-utilities\") pod \"08b1e05d-7b30-4e51-83c3-90892198f91d\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.627186 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-catalog-content\") pod \"08b1e05d-7b30-4e51-83c3-90892198f91d\" (UID: \"08b1e05d-7b30-4e51-83c3-90892198f91d\") " Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.633150 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-utilities" (OuterVolumeSpecName: "utilities") pod "08b1e05d-7b30-4e51-83c3-90892198f91d" (UID: "08b1e05d-7b30-4e51-83c3-90892198f91d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.643916 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08b1e05d-7b30-4e51-83c3-90892198f91d-kube-api-access-l75hz" (OuterVolumeSpecName: "kube-api-access-l75hz") pod "08b1e05d-7b30-4e51-83c3-90892198f91d" (UID: "08b1e05d-7b30-4e51-83c3-90892198f91d"). InnerVolumeSpecName "kube-api-access-l75hz". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.666134 5109 generic.go:358] "Generic (PLEG): container finished" podID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerID="a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47" exitCode=0 Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.666288 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xdjv" event={"ID":"08b1e05d-7b30-4e51-83c3-90892198f91d","Type":"ContainerDied","Data":"a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47"} Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.666321 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7xdjv" event={"ID":"08b1e05d-7b30-4e51-83c3-90892198f91d","Type":"ContainerDied","Data":"50e7cd9cd1bf6352f7710ab27632b9f989108dff92a9bd9990a2cb4d9c6d5ef0"} Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.666342 5109 scope.go:117] "RemoveContainer" containerID="a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.666529 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7xdjv" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.670480 5109 generic.go:358] "Generic (PLEG): container finished" podID="a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596" containerID="b022ba5e9b0dd216bbf92cb8a69cac404c550c185d7710f0e161868cdb5ef67c" exitCode=0 Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.670583 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596","Type":"ContainerDied","Data":"b022ba5e9b0dd216bbf92cb8a69cac404c550c185d7710f0e161868cdb5ef67c"} Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.672755 5109 generic.go:358] "Generic (PLEG): container finished" podID="1751cd13-4bda-4f90-9998-306a003b25ab" containerID="7ab02dd030d9048ebe6fca705655d45e4b4fae25dbcc7d1120e76c1716a53c07" exitCode=0 Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.673637 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdpj" event={"ID":"1751cd13-4bda-4f90-9998-306a003b25ab","Type":"ContainerDied","Data":"7ab02dd030d9048ebe6fca705655d45e4b4fae25dbcc7d1120e76c1716a53c07"} Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.681045 5109 scope.go:117] "RemoveContainer" containerID="60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.682324 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08b1e05d-7b30-4e51-83c3-90892198f91d" (UID: "08b1e05d-7b30-4e51-83c3-90892198f91d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.701072 5109 scope.go:117] "RemoveContainer" containerID="c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.719524 5109 scope.go:117] "RemoveContainer" containerID="a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47" Dec 11 16:55:38 crc kubenswrapper[5109]: E1211 16:55:38.720019 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47\": container with ID starting with a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47 not found: ID does not exist" containerID="a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.720049 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47"} err="failed to get container status \"a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47\": rpc error: code = NotFound desc = could not find container \"a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47\": container with ID starting with a91c01b36e2373c2650cf36985b37817cdb4d70dc68b7ed90559e5b28469ea47 not found: ID does not exist" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.720087 5109 scope.go:117] "RemoveContainer" containerID="60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f" Dec 11 16:55:38 crc kubenswrapper[5109]: E1211 16:55:38.720343 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f\": container with ID starting with 60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f not found: ID does not exist" containerID="60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.720364 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f"} err="failed to get container status \"60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f\": rpc error: code = NotFound desc = could not find container \"60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f\": container with ID starting with 60cba6d312c17fee539d627c6480a03f7aa0defeca0d2206add277798cd7351f not found: ID does not exist" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.720382 5109 scope.go:117] "RemoveContainer" containerID="c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0" Dec 11 16:55:38 crc kubenswrapper[5109]: E1211 16:55:38.720687 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0\": container with ID starting with c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0 not found: ID does not exist" containerID="c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.720710 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0"} err="failed to get container status \"c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0\": rpc error: code = NotFound desc = could not find container \"c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0\": container with ID starting with c30fece17d9ae85d968b31152367cd69ee3f82787aa3d092dc34d5c71a692cd0 not found: ID does not exist" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.728372 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-l75hz\" (UniqueName: \"kubernetes.io/projected/08b1e05d-7b30-4e51-83c3-90892198f91d-kube-api-access-l75hz\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.728402 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.728413 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08b1e05d-7b30-4e51-83c3-90892198f91d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.989154 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7xdjv"] Dec 11 16:55:38 crc kubenswrapper[5109]: I1211 16:55:38.989394 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7xdjv"] Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.388399 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.437470 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-catalog-content\") pod \"1751cd13-4bda-4f90-9998-306a003b25ab\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.437560 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhh5z\" (UniqueName: \"kubernetes.io/projected/1751cd13-4bda-4f90-9998-306a003b25ab-kube-api-access-hhh5z\") pod \"1751cd13-4bda-4f90-9998-306a003b25ab\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.437636 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-utilities\") pod \"1751cd13-4bda-4f90-9998-306a003b25ab\" (UID: \"1751cd13-4bda-4f90-9998-306a003b25ab\") " Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.439083 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-utilities" (OuterVolumeSpecName: "utilities") pod "1751cd13-4bda-4f90-9998-306a003b25ab" (UID: "1751cd13-4bda-4f90-9998-306a003b25ab"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.443426 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1751cd13-4bda-4f90-9998-306a003b25ab-kube-api-access-hhh5z" (OuterVolumeSpecName: "kube-api-access-hhh5z") pod "1751cd13-4bda-4f90-9998-306a003b25ab" (UID: "1751cd13-4bda-4f90-9998-306a003b25ab"). InnerVolumeSpecName "kube-api-access-hhh5z". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.489356 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1751cd13-4bda-4f90-9998-306a003b25ab" (UID: "1751cd13-4bda-4f90-9998-306a003b25ab"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.538840 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.538869 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1751cd13-4bda-4f90-9998-306a003b25ab-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.538881 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hhh5z\" (UniqueName: \"kubernetes.io/projected/1751cd13-4bda-4f90-9998-306a003b25ab-kube-api-access-hhh5z\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.589765 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590275 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="extract-utilities" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590286 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="extract-utilities" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590307 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="extract-utilities" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590312 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="extract-utilities" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590323 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="extract-content" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590329 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="extract-content" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590342 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="registry-server" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590347 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="registry-server" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590354 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="extract-content" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590359 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="extract-content" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590365 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="registry-server" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590370 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="registry-server" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590450 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" containerName="registry-server" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.590460 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" containerName="registry-server" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.601864 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.602577 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.681901 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdpj" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.681905 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdpj" event={"ID":"1751cd13-4bda-4f90-9998-306a003b25ab","Type":"ContainerDied","Data":"fdb07aec719d711bdfc9cd03bab02d09c95c019aed94c616f93b26c870557b04"} Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.682211 5109 scope.go:117] "RemoveContainer" containerID="7ab02dd030d9048ebe6fca705655d45e4b4fae25dbcc7d1120e76c1716a53c07" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.699479 5109 scope.go:117] "RemoveContainer" containerID="993e1fb3073fafbdf1263a62bc088c890ac0af75bd20b18da26e41f166b856bc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.716868 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5pdpj"] Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.719905 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5pdpj"] Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.739161 5109 scope.go:117] "RemoveContainer" containerID="f6c6f7cd694f658ee619f679e87a27ddc0482807988ce8aea450047ffe134d27" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.741264 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-kubelet-dir\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.741429 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-var-lock\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.741516 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4fc4ad8-69ad-4903-9936-b4936f5483af-kube-api-access\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.842942 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-var-lock\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.843171 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4fc4ad8-69ad-4903-9936-b4936f5483af-kube-api-access\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.843243 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-kubelet-dir\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.843341 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-var-lock\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.843626 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-kubelet-dir\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.860159 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4fc4ad8-69ad-4903-9936-b4936f5483af-kube-api-access\") pod \"installer-12-crc\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.882345 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.918904 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.938809 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfq9w"] Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.942291 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pfq9w" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="registry-server" containerID="cri-o://a7c8e710f96d5b14dec9182f53e41a00dff8a342d023fab570334dcb8dbcbc85" gracePeriod=2 Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.949195 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kubelet-dir\") pod \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.949304 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596" (UID: "a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.949333 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kube-api-access\") pod \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\" (UID: \"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596\") " Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.950040 5109 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:39 crc kubenswrapper[5109]: I1211 16:55:39.952506 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596" (UID: "a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.051896 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.091318 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-12-crc"] Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.139610 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrqhg"] Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.566153 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nrqhg" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="registry-server" containerID="cri-o://56272c96d2cd188288af93f0f82c691ba3c48eb076b1c089bbea32cc0a13ef22" gracePeriod=2 Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.693361 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-12-crc" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.693616 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-12-crc" event={"ID":"a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596","Type":"ContainerDied","Data":"0c29a23f300c3d13a154582ddb468322eac8be42d61ac24e221d9dd34d6b852a"} Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.693666 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c29a23f300c3d13a154582ddb468322eac8be42d61ac24e221d9dd34d6b852a" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.704020 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"b4fc4ad8-69ad-4903-9936-b4936f5483af","Type":"ContainerStarted","Data":"068dcea8d8f6435901cd43b3699b58a87e17b89edc70c4446c11fbbf66b22238"} Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.709456 5109 generic.go:358] "Generic (PLEG): container finished" podID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerID="a7c8e710f96d5b14dec9182f53e41a00dff8a342d023fab570334dcb8dbcbc85" exitCode=0 Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.709533 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfq9w" event={"ID":"634b1a50-cd7a-4673-b3bd-d49ab2a1114e","Type":"ContainerDied","Data":"a7c8e710f96d5b14dec9182f53e41a00dff8a342d023fab570334dcb8dbcbc85"} Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.712817 5109 generic.go:358] "Generic (PLEG): container finished" podID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerID="56272c96d2cd188288af93f0f82c691ba3c48eb076b1c089bbea32cc0a13ef22" exitCode=0 Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.712910 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrqhg" event={"ID":"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d","Type":"ContainerDied","Data":"56272c96d2cd188288af93f0f82c691ba3c48eb076b1c089bbea32cc0a13ef22"} Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.892997 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.912820 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08b1e05d-7b30-4e51-83c3-90892198f91d" path="/var/lib/kubelet/pods/08b1e05d-7b30-4e51-83c3-90892198f91d/volumes" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.913386 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1751cd13-4bda-4f90-9998-306a003b25ab" path="/var/lib/kubelet/pods/1751cd13-4bda-4f90-9998-306a003b25ab/volumes" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.976254 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-utilities\") pod \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.976326 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n26jq\" (UniqueName: \"kubernetes.io/projected/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-kube-api-access-n26jq\") pod \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.976373 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-catalog-content\") pod \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\" (UID: \"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d\") " Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.977199 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-utilities" (OuterVolumeSpecName: "utilities") pod "7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" (UID: "7c4ce6e5-c194-4c26-a2c1-d61a0542af3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:40 crc kubenswrapper[5109]: I1211 16:55:40.985378 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-kube-api-access-n26jq" (OuterVolumeSpecName: "kube-api-access-n26jq") pod "7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" (UID: "7c4ce6e5-c194-4c26-a2c1-d61a0542af3d"). InnerVolumeSpecName "kube-api-access-n26jq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.078233 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.078269 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-n26jq\" (UniqueName: \"kubernetes.io/projected/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-kube-api-access-n26jq\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.091012 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" (UID: "7c4ce6e5-c194-4c26-a2c1-d61a0542af3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.179316 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.371587 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.482847 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjc6q\" (UniqueName: \"kubernetes.io/projected/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-kube-api-access-fjc6q\") pod \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.482910 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-utilities\") pod \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.483034 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-catalog-content\") pod \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\" (UID: \"634b1a50-cd7a-4673-b3bd-d49ab2a1114e\") " Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.483955 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-utilities" (OuterVolumeSpecName: "utilities") pod "634b1a50-cd7a-4673-b3bd-d49ab2a1114e" (UID: "634b1a50-cd7a-4673-b3bd-d49ab2a1114e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.486400 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.486780 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-kube-api-access-fjc6q" (OuterVolumeSpecName: "kube-api-access-fjc6q") pod "634b1a50-cd7a-4673-b3bd-d49ab2a1114e" (UID: "634b1a50-cd7a-4673-b3bd-d49ab2a1114e"). InnerVolumeSpecName "kube-api-access-fjc6q". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.493971 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "634b1a50-cd7a-4673-b3bd-d49ab2a1114e" (UID: "634b1a50-cd7a-4673-b3bd-d49ab2a1114e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.587577 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-fjc6q\" (UniqueName: \"kubernetes.io/projected/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-kube-api-access-fjc6q\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.587617 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/634b1a50-cd7a-4673-b3bd-d49ab2a1114e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.718784 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"b4fc4ad8-69ad-4903-9936-b4936f5483af","Type":"ContainerStarted","Data":"bcc6951db5af30060d139dbaf039de6eeeedd437f251f8b24bbd118df45187e0"} Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.720761 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pfq9w" event={"ID":"634b1a50-cd7a-4673-b3bd-d49ab2a1114e","Type":"ContainerDied","Data":"6cb6e2f2b9495301a2c3f57b150f852269d71e7daa41b3728ff058ac9e511797"} Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.720809 5109 scope.go:117] "RemoveContainer" containerID="a7c8e710f96d5b14dec9182f53e41a00dff8a342d023fab570334dcb8dbcbc85" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.720818 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pfq9w" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.724215 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nrqhg" event={"ID":"7c4ce6e5-c194-4c26-a2c1-d61a0542af3d","Type":"ContainerDied","Data":"9b0a768865f60fdbef198c4ad661f8ae9dac10331aa5c14173114dfde4cb4cc1"} Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.724322 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nrqhg" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.736344 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-12-crc" podStartSLOduration=2.73632417 podStartE2EDuration="2.73632417s" podCreationTimestamp="2025-12-11 16:55:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:41.735668304 +0000 UTC m=+179.415359770" watchObservedRunningTime="2025-12-11 16:55:41.73632417 +0000 UTC m=+179.416015636" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.745409 5109 scope.go:117] "RemoveContainer" containerID="7db66df61f0fb5f9d15de57bab0bfc3c6b1a5330ee9019cb4cee469a50f5d6b0" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.768948 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nrqhg"] Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.768995 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nrqhg"] Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.779913 5109 scope.go:117] "RemoveContainer" containerID="024ec6d69b690d4496e3b3f5cc484934437f249e9f7fc9bb1cfb2b438166833c" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.811593 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfq9w"] Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.812435 5109 scope.go:117] "RemoveContainer" containerID="56272c96d2cd188288af93f0f82c691ba3c48eb076b1c089bbea32cc0a13ef22" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.816591 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pfq9w"] Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.843857 5109 scope.go:117] "RemoveContainer" containerID="e63fdec881e3ce29accee172651a4f84892cb8c49c06932ec600cbbf85dc4821" Dec 11 16:55:41 crc kubenswrapper[5109]: I1211 16:55:41.867117 5109 scope.go:117] "RemoveContainer" containerID="92d03fccc279074df8d380ec9372f9dbf494c9dcde67692c039dd26e5ed56ee6" Dec 11 16:55:42 crc kubenswrapper[5109]: I1211 16:55:42.905880 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" path="/var/lib/kubelet/pods/634b1a50-cd7a-4673-b3bd-d49ab2a1114e/volumes" Dec 11 16:55:42 crc kubenswrapper[5109]: I1211 16:55:42.906642 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" path="/var/lib/kubelet/pods/7c4ce6e5-c194-4c26-a2c1-d61a0542af3d/volumes" Dec 11 16:55:48 crc kubenswrapper[5109]: I1211 16:55:48.957141 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5dd796b77f-l9zgm"] Dec 11 16:55:48 crc kubenswrapper[5109]: I1211 16:55:48.957948 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" podUID="06ae51de-2665-44c0-8a42-bfc7bb42e7d8" containerName="controller-manager" containerID="cri-o://f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23" gracePeriod=30 Dec 11 16:55:48 crc kubenswrapper[5109]: I1211 16:55:48.978683 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d"] Dec 11 16:55:48 crc kubenswrapper[5109]: I1211 16:55:48.978953 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" podUID="b27c4651-d447-4fd1-8f8d-fc81e9778230" containerName="route-controller-manager" containerID="cri-o://90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d" gracePeriod=30 Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.297150 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.330306 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331049 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596" containerName="pruner" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331071 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596" containerName="pruner" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331087 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="extract-content" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331096 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="extract-content" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331108 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="registry-server" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331114 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="registry-server" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331132 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b27c4651-d447-4fd1-8f8d-fc81e9778230" containerName="route-controller-manager" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331139 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="b27c4651-d447-4fd1-8f8d-fc81e9778230" containerName="route-controller-manager" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331146 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="registry-server" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331152 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="registry-server" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331162 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="extract-utilities" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331170 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="extract-utilities" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331184 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="extract-utilities" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331192 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="extract-utilities" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331200 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="extract-content" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331208 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="extract-content" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331327 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="a7ef01bd-dacd-4a7a-a8a4-cc5222bc5596" containerName="pruner" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331342 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="7c4ce6e5-c194-4c26-a2c1-d61a0542af3d" containerName="registry-server" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331350 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="634b1a50-cd7a-4673-b3bd-d49ab2a1114e" containerName="registry-server" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.331363 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="b27c4651-d447-4fd1-8f8d-fc81e9778230" containerName="route-controller-manager" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.335814 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.344599 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.419497 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-client-ca\") pod \"b27c4651-d447-4fd1-8f8d-fc81e9778230\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.419667 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-config\") pod \"b27c4651-d447-4fd1-8f8d-fc81e9778230\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.419752 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b27c4651-d447-4fd1-8f8d-fc81e9778230-serving-cert\") pod \"b27c4651-d447-4fd1-8f8d-fc81e9778230\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.419823 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvsx8\" (UniqueName: \"kubernetes.io/projected/b27c4651-d447-4fd1-8f8d-fc81e9778230-kube-api-access-rvsx8\") pod \"b27c4651-d447-4fd1-8f8d-fc81e9778230\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.419874 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b27c4651-d447-4fd1-8f8d-fc81e9778230-tmp\") pod \"b27c4651-d447-4fd1-8f8d-fc81e9778230\" (UID: \"b27c4651-d447-4fd1-8f8d-fc81e9778230\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.420403 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b27c4651-d447-4fd1-8f8d-fc81e9778230-tmp" (OuterVolumeSpecName: "tmp") pod "b27c4651-d447-4fd1-8f8d-fc81e9778230" (UID: "b27c4651-d447-4fd1-8f8d-fc81e9778230"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.420572 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-client-ca" (OuterVolumeSpecName: "client-ca") pod "b27c4651-d447-4fd1-8f8d-fc81e9778230" (UID: "b27c4651-d447-4fd1-8f8d-fc81e9778230"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.420730 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-config" (OuterVolumeSpecName: "config") pod "b27c4651-d447-4fd1-8f8d-fc81e9778230" (UID: "b27c4651-d447-4fd1-8f8d-fc81e9778230"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.426655 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b27c4651-d447-4fd1-8f8d-fc81e9778230-kube-api-access-rvsx8" (OuterVolumeSpecName: "kube-api-access-rvsx8") pod "b27c4651-d447-4fd1-8f8d-fc81e9778230" (UID: "b27c4651-d447-4fd1-8f8d-fc81e9778230"). InnerVolumeSpecName "kube-api-access-rvsx8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.427016 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b27c4651-d447-4fd1-8f8d-fc81e9778230-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b27c4651-d447-4fd1-8f8d-fc81e9778230" (UID: "b27c4651-d447-4fd1-8f8d-fc81e9778230"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.521531 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7qx4\" (UniqueName: \"kubernetes.io/projected/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-kube-api-access-g7qx4\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.521606 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-tmp\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.521711 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-serving-cert\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.521809 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-client-ca\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.521871 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-config\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.522093 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rvsx8\" (UniqueName: \"kubernetes.io/projected/b27c4651-d447-4fd1-8f8d-fc81e9778230-kube-api-access-rvsx8\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.522126 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/b27c4651-d447-4fd1-8f8d-fc81e9778230-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.522139 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.522151 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b27c4651-d447-4fd1-8f8d-fc81e9778230-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.522160 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b27c4651-d447-4fd1-8f8d-fc81e9778230-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.592378 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.621535 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f5d8d6576-glxqf"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.622245 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="06ae51de-2665-44c0-8a42-bfc7bb42e7d8" containerName="controller-manager" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.622268 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="06ae51de-2665-44c0-8a42-bfc7bb42e7d8" containerName="controller-manager" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.622438 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="06ae51de-2665-44c0-8a42-bfc7bb42e7d8" containerName="controller-manager" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.623528 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-client-ca\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.623696 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-config\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.623783 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-g7qx4\" (UniqueName: \"kubernetes.io/projected/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-kube-api-access-g7qx4\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.623828 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-tmp\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.623861 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-serving-cert\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.624638 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-client-ca\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.624696 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-config\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.624988 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-tmp\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.630089 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-serving-cert\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.645572 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7qx4\" (UniqueName: \"kubernetes.io/projected/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-kube-api-access-g7qx4\") pod \"route-controller-manager-fd8bc5f79-b4mzd\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.646919 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f5d8d6576-glxqf"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.647373 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.652505 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.724595 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-client-ca\") pod \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.724779 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-proxy-ca-bundles\") pod \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.724814 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-serving-cert\") pod \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.724834 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-tmp\") pod \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.724881 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm85n\" (UniqueName: \"kubernetes.io/projected/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-kube-api-access-lm85n\") pod \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.724902 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-config\") pod \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\" (UID: \"06ae51de-2665-44c0-8a42-bfc7bb42e7d8\") " Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.725958 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-config" (OuterVolumeSpecName: "config") pod "06ae51de-2665-44c0-8a42-bfc7bb42e7d8" (UID: "06ae51de-2665-44c0-8a42-bfc7bb42e7d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.726002 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "06ae51de-2665-44c0-8a42-bfc7bb42e7d8" (UID: "06ae51de-2665-44c0-8a42-bfc7bb42e7d8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.726000 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-client-ca" (OuterVolumeSpecName: "client-ca") pod "06ae51de-2665-44c0-8a42-bfc7bb42e7d8" (UID: "06ae51de-2665-44c0-8a42-bfc7bb42e7d8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.726108 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-tmp" (OuterVolumeSpecName: "tmp") pod "06ae51de-2665-44c0-8a42-bfc7bb42e7d8" (UID: "06ae51de-2665-44c0-8a42-bfc7bb42e7d8"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.729230 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-kube-api-access-lm85n" (OuterVolumeSpecName: "kube-api-access-lm85n") pod "06ae51de-2665-44c0-8a42-bfc7bb42e7d8" (UID: "06ae51de-2665-44c0-8a42-bfc7bb42e7d8"). InnerVolumeSpecName "kube-api-access-lm85n". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.729845 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "06ae51de-2665-44c0-8a42-bfc7bb42e7d8" (UID: "06ae51de-2665-44c0-8a42-bfc7bb42e7d8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.783394 5109 generic.go:358] "Generic (PLEG): container finished" podID="b27c4651-d447-4fd1-8f8d-fc81e9778230" containerID="90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d" exitCode=0 Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.783440 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" event={"ID":"b27c4651-d447-4fd1-8f8d-fc81e9778230","Type":"ContainerDied","Data":"90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d"} Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.783484 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" event={"ID":"b27c4651-d447-4fd1-8f8d-fc81e9778230","Type":"ContainerDied","Data":"48b7e7ef3661efa32fdd3b9f1ae0d65f97ae5ae2f1c95abd1ac21d02ecad02a0"} Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.783511 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.783529 5109 scope.go:117] "RemoveContainer" containerID="90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.786704 5109 generic.go:358] "Generic (PLEG): container finished" podID="06ae51de-2665-44c0-8a42-bfc7bb42e7d8" containerID="f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23" exitCode=0 Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.786865 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.786863 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" event={"ID":"06ae51de-2665-44c0-8a42-bfc7bb42e7d8","Type":"ContainerDied","Data":"f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23"} Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.787097 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5dd796b77f-l9zgm" event={"ID":"06ae51de-2665-44c0-8a42-bfc7bb42e7d8","Type":"ContainerDied","Data":"1e00f37db6a65a8ad705e78dc17a1b2e863b5a4b0b47a1fcc09df103c59a7239"} Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.809251 5109 scope.go:117] "RemoveContainer" containerID="90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d" Dec 11 16:55:51 crc kubenswrapper[5109]: E1211 16:55:51.809785 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d\": container with ID starting with 90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d not found: ID does not exist" containerID="90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.809834 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d"} err="failed to get container status \"90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d\": rpc error: code = NotFound desc = could not find container \"90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d\": container with ID starting with 90ad3647756db0ad404686aa2d96384cb5e6cbf9a34763c501e8f4c31a0d4a9d not found: ID does not exist" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.809890 5109 scope.go:117] "RemoveContainer" containerID="f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.833484 5109 scope.go:117] "RemoveContainer" containerID="f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.833816 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.834962 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-config\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835013 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-proxy-ca-bundles\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835043 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/90f5523e-637c-4625-81cc-52c77ebff6f1-tmp\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835095 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn7f8\" (UniqueName: \"kubernetes.io/projected/90f5523e-637c-4625-81cc-52c77ebff6f1-kube-api-access-gn7f8\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835138 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90f5523e-637c-4625-81cc-52c77ebff6f1-serving-cert\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835284 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-client-ca\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835345 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835361 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835373 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835385 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lm85n\" (UniqueName: \"kubernetes.io/projected/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-kube-api-access-lm85n\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835400 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.835413 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/06ae51de-2665-44c0-8a42-bfc7bb42e7d8-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:55:51 crc kubenswrapper[5109]: E1211 16:55:51.838775 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23\": container with ID starting with f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23 not found: ID does not exist" containerID="f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.838825 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23"} err="failed to get container status \"f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23\": rpc error: code = NotFound desc = could not find container \"f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23\": container with ID starting with f32bf769878a72c5f11748735bab606645d518452155dab8dee51d9dfeca4d23 not found: ID does not exist" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.842667 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-8997cb6f7-6rt4d"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.850307 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5dd796b77f-l9zgm"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.852491 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd"] Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.856308 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5dd796b77f-l9zgm"] Dec 11 16:55:51 crc kubenswrapper[5109]: W1211 16:55:51.857578 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0d5c3b7_c2d7_4b5d_93a2_2b7f4b5019c5.slice/crio-caa2bd06412a6925c14162f6bfae39fc9028c5f00f30784951321f8a813df8a0 WatchSource:0}: Error finding container caa2bd06412a6925c14162f6bfae39fc9028c5f00f30784951321f8a813df8a0: Status 404 returned error can't find the container with id caa2bd06412a6925c14162f6bfae39fc9028c5f00f30784951321f8a813df8a0 Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.936100 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-gn7f8\" (UniqueName: \"kubernetes.io/projected/90f5523e-637c-4625-81cc-52c77ebff6f1-kube-api-access-gn7f8\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.936158 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90f5523e-637c-4625-81cc-52c77ebff6f1-serving-cert\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.936238 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-client-ca\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.936301 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-config\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.936322 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-proxy-ca-bundles\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.936348 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/90f5523e-637c-4625-81cc-52c77ebff6f1-tmp\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.936899 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/90f5523e-637c-4625-81cc-52c77ebff6f1-tmp\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.938646 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-config\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.942288 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-proxy-ca-bundles\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.942511 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-client-ca\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.944740 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90f5523e-637c-4625-81cc-52c77ebff6f1-serving-cert\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.955815 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn7f8\" (UniqueName: \"kubernetes.io/projected/90f5523e-637c-4625-81cc-52c77ebff6f1-kube-api-access-gn7f8\") pod \"controller-manager-f5d8d6576-glxqf\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:51 crc kubenswrapper[5109]: I1211 16:55:51.968897 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:52 crc kubenswrapper[5109]: W1211 16:55:52.159214 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod90f5523e_637c_4625_81cc_52c77ebff6f1.slice/crio-2264cbcae01fc7bbb80980b5c0ecb6c239dba7c5c17bcadfba8ad67031a1cc50 WatchSource:0}: Error finding container 2264cbcae01fc7bbb80980b5c0ecb6c239dba7c5c17bcadfba8ad67031a1cc50: Status 404 returned error can't find the container with id 2264cbcae01fc7bbb80980b5c0ecb6c239dba7c5c17bcadfba8ad67031a1cc50 Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.164380 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f5d8d6576-glxqf"] Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.796285 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" event={"ID":"90f5523e-637c-4625-81cc-52c77ebff6f1","Type":"ContainerStarted","Data":"02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1"} Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.796646 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" event={"ID":"90f5523e-637c-4625-81cc-52c77ebff6f1","Type":"ContainerStarted","Data":"2264cbcae01fc7bbb80980b5c0ecb6c239dba7c5c17bcadfba8ad67031a1cc50"} Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.796669 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.799352 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" event={"ID":"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5","Type":"ContainerStarted","Data":"3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649"} Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.799395 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" event={"ID":"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5","Type":"ContainerStarted","Data":"caa2bd06412a6925c14162f6bfae39fc9028c5f00f30784951321f8a813df8a0"} Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.799599 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.815469 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" podStartSLOduration=4.815444739 podStartE2EDuration="4.815444739s" podCreationTimestamp="2025-12-11 16:55:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:52.809812949 +0000 UTC m=+190.489504415" watchObservedRunningTime="2025-12-11 16:55:52.815444739 +0000 UTC m=+190.495136225" Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.925625 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="06ae51de-2665-44c0-8a42-bfc7bb42e7d8" path="/var/lib/kubelet/pods/06ae51de-2665-44c0-8a42-bfc7bb42e7d8/volumes" Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.926405 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b27c4651-d447-4fd1-8f8d-fc81e9778230" path="/var/lib/kubelet/pods/b27c4651-d447-4fd1-8f8d-fc81e9778230/volumes" Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.935927 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:55:52 crc kubenswrapper[5109]: I1211 16:55:52.951024 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" podStartSLOduration=3.951007765 podStartE2EDuration="3.951007765s" podCreationTimestamp="2025-12-11 16:55:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:55:52.829387545 +0000 UTC m=+190.509079011" watchObservedRunningTime="2025-12-11 16:55:52.951007765 +0000 UTC m=+190.630699231" Dec 11 16:55:53 crc kubenswrapper[5109]: I1211 16:55:53.348704 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:55:59 crc kubenswrapper[5109]: I1211 16:55:59.515880 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-fhkjl" Dec 11 16:56:03 crc kubenswrapper[5109]: I1211 16:56:03.508578 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mr25z"] Dec 11 16:56:08 crc kubenswrapper[5109]: I1211 16:56:08.995565 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f5d8d6576-glxqf"] Dec 11 16:56:08 crc kubenswrapper[5109]: I1211 16:56:08.996304 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" podUID="90f5523e-637c-4625-81cc-52c77ebff6f1" containerName="controller-manager" containerID="cri-o://02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1" gracePeriod=30 Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.020037 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.020633 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" podUID="d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" containerName="route-controller-manager" containerID="cri-o://3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649" gracePeriod=30 Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.497111 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.521913 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.522924 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" containerName="route-controller-manager" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.523045 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" containerName="route-controller-manager" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.523255 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" containerName="route-controller-manager" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.528635 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.534955 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.582417 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-config\") pod \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.582492 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7qx4\" (UniqueName: \"kubernetes.io/projected/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-kube-api-access-g7qx4\") pod \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.582540 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-tmp\") pod \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.582570 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-client-ca\") pod \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.582604 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-serving-cert\") pod \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\" (UID: \"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.583706 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-tmp" (OuterVolumeSpecName: "tmp") pod "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" (UID: "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.584085 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-client-ca" (OuterVolumeSpecName: "client-ca") pod "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" (UID: "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.584103 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-config" (OuterVolumeSpecName: "config") pod "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" (UID: "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.588224 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-kube-api-access-g7qx4" (OuterVolumeSpecName: "kube-api-access-g7qx4") pod "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" (UID: "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5"). InnerVolumeSpecName "kube-api-access-g7qx4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.588350 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" (UID: "d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.683935 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-client-ca\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684224 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-config\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684278 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-serving-cert\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684316 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-tmp\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684345 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2chqp\" (UniqueName: \"kubernetes.io/projected/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-kube-api-access-2chqp\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684418 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684434 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684446 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684462 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.684473 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-g7qx4\" (UniqueName: \"kubernetes.io/projected/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5-kube-api-access-g7qx4\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.687715 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.717663 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-b867f88c9-pq5kk"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.718410 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="90f5523e-637c-4625-81cc-52c77ebff6f1" containerName="controller-manager" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.718434 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="90f5523e-637c-4625-81cc-52c77ebff6f1" containerName="controller-manager" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.718566 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="90f5523e-637c-4625-81cc-52c77ebff6f1" containerName="controller-manager" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.725159 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.731329 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b867f88c9-pq5kk"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785010 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn7f8\" (UniqueName: \"kubernetes.io/projected/90f5523e-637c-4625-81cc-52c77ebff6f1-kube-api-access-gn7f8\") pod \"90f5523e-637c-4625-81cc-52c77ebff6f1\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785107 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-client-ca\") pod \"90f5523e-637c-4625-81cc-52c77ebff6f1\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785138 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-config\") pod \"90f5523e-637c-4625-81cc-52c77ebff6f1\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785173 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90f5523e-637c-4625-81cc-52c77ebff6f1-serving-cert\") pod \"90f5523e-637c-4625-81cc-52c77ebff6f1\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785204 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/90f5523e-637c-4625-81cc-52c77ebff6f1-tmp\") pod \"90f5523e-637c-4625-81cc-52c77ebff6f1\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785232 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-proxy-ca-bundles\") pod \"90f5523e-637c-4625-81cc-52c77ebff6f1\" (UID: \"90f5523e-637c-4625-81cc-52c77ebff6f1\") " Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785352 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-client-ca\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785406 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-config\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785459 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-serving-cert\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785495 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-tmp\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.785516 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-2chqp\" (UniqueName: \"kubernetes.io/projected/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-kube-api-access-2chqp\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.786515 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-client-ca" (OuterVolumeSpecName: "client-ca") pod "90f5523e-637c-4625-81cc-52c77ebff6f1" (UID: "90f5523e-637c-4625-81cc-52c77ebff6f1"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.786542 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "90f5523e-637c-4625-81cc-52c77ebff6f1" (UID: "90f5523e-637c-4625-81cc-52c77ebff6f1"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.786607 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-config" (OuterVolumeSpecName: "config") pod "90f5523e-637c-4625-81cc-52c77ebff6f1" (UID: "90f5523e-637c-4625-81cc-52c77ebff6f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.786675 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-tmp\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.787035 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-client-ca\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.787381 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90f5523e-637c-4625-81cc-52c77ebff6f1-tmp" (OuterVolumeSpecName: "tmp") pod "90f5523e-637c-4625-81cc-52c77ebff6f1" (UID: "90f5523e-637c-4625-81cc-52c77ebff6f1"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.788027 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-config\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.789227 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90f5523e-637c-4625-81cc-52c77ebff6f1-kube-api-access-gn7f8" (OuterVolumeSpecName: "kube-api-access-gn7f8") pod "90f5523e-637c-4625-81cc-52c77ebff6f1" (UID: "90f5523e-637c-4625-81cc-52c77ebff6f1"). InnerVolumeSpecName "kube-api-access-gn7f8". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.789936 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-serving-cert\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.789956 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/90f5523e-637c-4625-81cc-52c77ebff6f1-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "90f5523e-637c-4625-81cc-52c77ebff6f1" (UID: "90f5523e-637c-4625-81cc-52c77ebff6f1"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.801638 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-2chqp\" (UniqueName: \"kubernetes.io/projected/20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21-kube-api-access-2chqp\") pod \"route-controller-manager-57447f659f-6gqfr\" (UID: \"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21\") " pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.844143 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887055 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f614b7ea-57c0-4317-a885-175a6b078602-tmp\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887114 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f614b7ea-57c0-4317-a885-175a6b078602-serving-cert\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887140 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-proxy-ca-bundles\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887193 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk7q4\" (UniqueName: \"kubernetes.io/projected/f614b7ea-57c0-4317-a885-175a6b078602-kube-api-access-xk7q4\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887237 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-client-ca\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887262 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-config\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887362 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-gn7f8\" (UniqueName: \"kubernetes.io/projected/90f5523e-637c-4625-81cc-52c77ebff6f1-kube-api-access-gn7f8\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887379 5109 reconciler_common.go:299] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-client-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887392 5109 reconciler_common.go:299] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-config\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887403 5109 reconciler_common.go:299] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/90f5523e-637c-4625-81cc-52c77ebff6f1-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887413 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/90f5523e-637c-4625-81cc-52c77ebff6f1-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.887424 5109 reconciler_common.go:299] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/90f5523e-637c-4625-81cc-52c77ebff6f1-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.926371 5109 generic.go:358] "Generic (PLEG): container finished" podID="90f5523e-637c-4625-81cc-52c77ebff6f1" containerID="02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1" exitCode=0 Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.926472 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" event={"ID":"90f5523e-637c-4625-81cc-52c77ebff6f1","Type":"ContainerDied","Data":"02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1"} Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.926512 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" event={"ID":"90f5523e-637c-4625-81cc-52c77ebff6f1","Type":"ContainerDied","Data":"2264cbcae01fc7bbb80980b5c0ecb6c239dba7c5c17bcadfba8ad67031a1cc50"} Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.926537 5109 scope.go:117] "RemoveContainer" containerID="02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.926719 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f5d8d6576-glxqf" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.938150 5109 generic.go:358] "Generic (PLEG): container finished" podID="d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" containerID="3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649" exitCode=0 Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.938331 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.938525 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" event={"ID":"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5","Type":"ContainerDied","Data":"3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649"} Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.938593 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd" event={"ID":"d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5","Type":"ContainerDied","Data":"caa2bd06412a6925c14162f6bfae39fc9028c5f00f30784951321f8a813df8a0"} Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.956383 5109 scope.go:117] "RemoveContainer" containerID="02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1" Dec 11 16:56:09 crc kubenswrapper[5109]: E1211 16:56:09.959193 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1\": container with ID starting with 02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1 not found: ID does not exist" containerID="02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.959230 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1"} err="failed to get container status \"02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1\": rpc error: code = NotFound desc = could not find container \"02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1\": container with ID starting with 02f26b1e26cb4be3ae6b99508f78c7edfb4deda945f108049a7b66c9c26e32e1 not found: ID does not exist" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.959254 5109 scope.go:117] "RemoveContainer" containerID="3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.973709 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.980913 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd8bc5f79-b4mzd"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.982695 5109 scope.go:117] "RemoveContainer" containerID="3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649" Dec 11 16:56:09 crc kubenswrapper[5109]: E1211 16:56:09.983348 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649\": container with ID starting with 3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649 not found: ID does not exist" containerID="3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.983678 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649"} err="failed to get container status \"3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649\": rpc error: code = NotFound desc = could not find container \"3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649\": container with ID starting with 3c02ef9ef5981066c0a801cc07f130b5ed61c16c89c72ff7c178292acb6cb649 not found: ID does not exist" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.985830 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f5d8d6576-glxqf"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.988666 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f614b7ea-57c0-4317-a885-175a6b078602-serving-cert\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.988698 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-proxy-ca-bundles\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.988756 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xk7q4\" (UniqueName: \"kubernetes.io/projected/f614b7ea-57c0-4317-a885-175a6b078602-kube-api-access-xk7q4\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.988779 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-client-ca\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.988795 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-config\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.988898 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f614b7ea-57c0-4317-a885-175a6b078602-tmp\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.989592 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-f5d8d6576-glxqf"] Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.991207 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-client-ca\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.991214 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-proxy-ca-bundles\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.992117 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f614b7ea-57c0-4317-a885-175a6b078602-config\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:09 crc kubenswrapper[5109]: I1211 16:56:09.992397 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f614b7ea-57c0-4317-a885-175a6b078602-tmp\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.001394 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f614b7ea-57c0-4317-a885-175a6b078602-serving-cert\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.015430 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk7q4\" (UniqueName: \"kubernetes.io/projected/f614b7ea-57c0-4317-a885-175a6b078602-kube-api-access-xk7q4\") pod \"controller-manager-b867f88c9-pq5kk\" (UID: \"f614b7ea-57c0-4317-a885-175a6b078602\") " pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.037647 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.214670 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-b867f88c9-pq5kk"] Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.246636 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr"] Dec 11 16:56:10 crc kubenswrapper[5109]: W1211 16:56:10.249867 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20a2b6ca_cd9a_4cc7_a38c_e7fc85074e21.slice/crio-87a4abbf8b859d2a99ab9c11ba1d2e249522acf80503ec811b27a2c284a2948c WatchSource:0}: Error finding container 87a4abbf8b859d2a99ab9c11ba1d2e249522acf80503ec811b27a2c284a2948c: Status 404 returned error can't find the container with id 87a4abbf8b859d2a99ab9c11ba1d2e249522acf80503ec811b27a2c284a2948c Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.908488 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90f5523e-637c-4625-81cc-52c77ebff6f1" path="/var/lib/kubelet/pods/90f5523e-637c-4625-81cc-52c77ebff6f1/volumes" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.909710 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5" path="/var/lib/kubelet/pods/d0d5c3b7-c2d7-4b5d-93a2-2b7f4b5019c5/volumes" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.945605 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" event={"ID":"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21","Type":"ContainerStarted","Data":"7fd4b7937c4ee5bfadc0a2b8fd0cccb7a0cf15b89b5841258f26a0b7b4bd9871"} Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.945649 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" event={"ID":"20a2b6ca-cd9a-4cc7-a38c-e7fc85074e21","Type":"ContainerStarted","Data":"87a4abbf8b859d2a99ab9c11ba1d2e249522acf80503ec811b27a2c284a2948c"} Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.945982 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.952483 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" event={"ID":"f614b7ea-57c0-4317-a885-175a6b078602","Type":"ContainerStarted","Data":"0978a376f082c288498628b8d589ed715b38a9a9c7727a50945841bd3e7a3e8f"} Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.952532 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" event={"ID":"f614b7ea-57c0-4317-a885-175a6b078602","Type":"ContainerStarted","Data":"5451b7b15399c1514eac6cfb5185732c204241cfb5549bd0f858da50e77b29c7"} Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.953098 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.960802 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.966209 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" podStartSLOduration=1.9661932549999999 podStartE2EDuration="1.966193255s" podCreationTimestamp="2025-12-11 16:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:56:10.964250081 +0000 UTC m=+208.643941557" watchObservedRunningTime="2025-12-11 16:56:10.966193255 +0000 UTC m=+208.645884721" Dec 11 16:56:10 crc kubenswrapper[5109]: I1211 16:56:10.984276 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-b867f88c9-pq5kk" podStartSLOduration=1.9842581799999999 podStartE2EDuration="1.98425818s" podCreationTimestamp="2025-12-11 16:56:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:56:10.981717548 +0000 UTC m=+208.661409034" watchObservedRunningTime="2025-12-11 16:56:10.98425818 +0000 UTC m=+208.663949646" Dec 11 16:56:11 crc kubenswrapper[5109]: I1211 16:56:11.528619 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-57447f659f-6gqfr" Dec 11 16:56:12 crc kubenswrapper[5109]: I1211 16:56:12.622604 5109 ???:1] "http: TLS handshake error from 192.168.126.11:57048: no serving certificate available for the kubelet" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.697860 5109 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754081 5109 kubelet.go:2547] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754156 5109 kubelet.go:2537] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754381 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754848 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" containerID="cri-o://680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690" gracePeriod=15 Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754892 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754909 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754922 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754930 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754944 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="setup" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754952 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="setup" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754966 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754973 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754981 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754990 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.754998 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755006 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755022 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755030 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755059 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755067 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755082 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" containerID="cri-o://beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036" gracePeriod=15 Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755053 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a" gracePeriod=15 Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755102 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" containerID="cri-o://45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9" gracePeriod=15 Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755148 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85" gracePeriod=15 Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755188 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755419 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755489 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755510 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-regeneration-controller" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755569 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755595 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755615 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755670 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-insecure-readyz" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.755687 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-cert-syncer" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.756187 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.756243 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.756274 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.756286 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a14caf222afb62aaabdc47808b6f944" containerName="kube-apiserver-check-endpoints" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.766349 5109 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="3a14caf222afb62aaabdc47808b6f944" podUID="57755cc5f99000cc11e193051474d4e2" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.798665 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.823519 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.823865 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824010 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824124 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824276 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824399 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824518 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824634 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824789 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.824947 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926456 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926504 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926531 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926568 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926599 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926619 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926635 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926668 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926683 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926700 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926777 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926815 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926834 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926854 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926873 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.926892 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/57755cc5f99000cc11e193051474d4e2-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.927296 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-ca-bundle-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.927326 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.927472 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:18 crc kubenswrapper[5109]: I1211 16:56:18.927484 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/57755cc5f99000cc11e193051474d4e2-tmp-dir\") pod \"kube-apiserver-crc\" (UID: \"57755cc5f99000cc11e193051474d4e2\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:19 crc kubenswrapper[5109]: I1211 16:56:19.017784 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-check-endpoints/3.log" Dec 11 16:56:19 crc kubenswrapper[5109]: I1211 16:56:19.019357 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Dec 11 16:56:19 crc kubenswrapper[5109]: I1211 16:56:19.020081 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036" exitCode=0 Dec 11 16:56:19 crc kubenswrapper[5109]: I1211 16:56:19.020115 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a" exitCode=0 Dec 11 16:56:19 crc kubenswrapper[5109]: I1211 16:56:19.020123 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85" exitCode=0 Dec 11 16:56:19 crc kubenswrapper[5109]: I1211 16:56:19.020132 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9" exitCode=2 Dec 11 16:56:19 crc kubenswrapper[5109]: I1211 16:56:19.020198 5109 scope.go:117] "RemoveContainer" containerID="6fb739ef3a81d36a2763bab1fbed470d8ad2bc7b6bb09353424ed941e1348f5d" Dec 11 16:56:20 crc kubenswrapper[5109]: I1211 16:56:20.036000 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.296954 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.298375 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.299292 5109 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356467 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356547 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356588 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356593 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356644 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356672 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") pod \"3a14caf222afb62aaabdc47808b6f944\" (UID: \"3a14caf222afb62aaabdc47808b6f944\") " Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356677 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.356682 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.357433 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir" (OuterVolumeSpecName: "ca-bundle-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "ca-bundle-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.357662 5109 reconciler_common.go:299] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.357704 5109 reconciler_common.go:299] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.357723 5109 reconciler_common.go:299] "Volume detached for volume \"ca-bundle-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-ca-bundle-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.357766 5109 reconciler_common.go:299] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/3a14caf222afb62aaabdc47808b6f944-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.360978 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir" (OuterVolumeSpecName: "tmp-dir") pod "3a14caf222afb62aaabdc47808b6f944" (UID: "3a14caf222afb62aaabdc47808b6f944"). InnerVolumeSpecName "tmp-dir". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:56:21 crc kubenswrapper[5109]: I1211 16:56:21.459762 5109 reconciler_common.go:299] "Volume detached for volume \"tmp-dir\" (UniqueName: \"kubernetes.io/empty-dir/3a14caf222afb62aaabdc47808b6f944-tmp-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.050561 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_3a14caf222afb62aaabdc47808b6f944/kube-apiserver-cert-syncer/0.log" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.052050 5109 generic.go:358] "Generic (PLEG): container finished" podID="3a14caf222afb62aaabdc47808b6f944" containerID="680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690" exitCode=0 Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.052174 5109 scope.go:117] "RemoveContainer" containerID="beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.052245 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.074029 5109 scope.go:117] "RemoveContainer" containerID="c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.083461 5109 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.097197 5109 scope.go:117] "RemoveContainer" containerID="f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.125682 5109 scope.go:117] "RemoveContainer" containerID="45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.152228 5109 scope.go:117] "RemoveContainer" containerID="680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.177773 5109 scope.go:117] "RemoveContainer" containerID="03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.242492 5109 scope.go:117] "RemoveContainer" containerID="beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036" Dec 11 16:56:22 crc kubenswrapper[5109]: E1211 16:56:22.242925 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036\": container with ID starting with beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036 not found: ID does not exist" containerID="beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.242970 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036"} err="failed to get container status \"beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036\": rpc error: code = NotFound desc = could not find container \"beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036\": container with ID starting with beff585dd86c01bc928276768289c63d13426ed57e6455b36fc2b2004a63a036 not found: ID does not exist" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.242998 5109 scope.go:117] "RemoveContainer" containerID="c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a" Dec 11 16:56:22 crc kubenswrapper[5109]: E1211 16:56:22.244537 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a\": container with ID starting with c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a not found: ID does not exist" containerID="c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.244581 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a"} err="failed to get container status \"c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a\": rpc error: code = NotFound desc = could not find container \"c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a\": container with ID starting with c4e11acc0f77eae5ec05ab618b44e13efd4ddc42eb38c64a908a8786131fb09a not found: ID does not exist" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.244603 5109 scope.go:117] "RemoveContainer" containerID="f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85" Dec 11 16:56:22 crc kubenswrapper[5109]: E1211 16:56:22.244976 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85\": container with ID starting with f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85 not found: ID does not exist" containerID="f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.245062 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85"} err="failed to get container status \"f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85\": rpc error: code = NotFound desc = could not find container \"f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85\": container with ID starting with f662d0c67bb5ea20c2b864f46077cc1033bc92a5e01bf266e7a372d62fbcbf85 not found: ID does not exist" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.245099 5109 scope.go:117] "RemoveContainer" containerID="45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9" Dec 11 16:56:22 crc kubenswrapper[5109]: E1211 16:56:22.245482 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9\": container with ID starting with 45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9 not found: ID does not exist" containerID="45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.245517 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9"} err="failed to get container status \"45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9\": rpc error: code = NotFound desc = could not find container \"45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9\": container with ID starting with 45f8a9edaf478fd872d2b522cb31fecb9e74beec77dd883a9f7120edd9f7c6a9 not found: ID does not exist" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.245540 5109 scope.go:117] "RemoveContainer" containerID="680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690" Dec 11 16:56:22 crc kubenswrapper[5109]: E1211 16:56:22.245815 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690\": container with ID starting with 680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690 not found: ID does not exist" containerID="680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.245875 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690"} err="failed to get container status \"680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690\": rpc error: code = NotFound desc = could not find container \"680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690\": container with ID starting with 680f9f1311d02dfc01a4a044d53aef3fbe33412337a9860c46cc950fb4059690 not found: ID does not exist" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.245894 5109 scope.go:117] "RemoveContainer" containerID="03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a" Dec 11 16:56:22 crc kubenswrapper[5109]: E1211 16:56:22.246269 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\": container with ID starting with 03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a not found: ID does not exist" containerID="03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.246304 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a"} err="failed to get container status \"03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\": rpc error: code = NotFound desc = could not find container \"03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a\": container with ID starting with 03fe1881a023832071a4d0955061f2313d32e05b6e33685a8306b2e908f8cc4a not found: ID does not exist" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.907929 5109 status_manager.go:895] "Failed to get status for pod" podUID="3a14caf222afb62aaabdc47808b6f944" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:22 crc kubenswrapper[5109]: I1211 16:56:22.914773 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a14caf222afb62aaabdc47808b6f944" path="/var/lib/kubelet/pods/3a14caf222afb62aaabdc47808b6f944/volumes" Dec 11 16:56:23 crc kubenswrapper[5109]: E1211 16:56:23.804255 5109 kubelet.go:3342] "Failed creating a mirror pod" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:23 crc kubenswrapper[5109]: I1211 16:56:23.804872 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:23 crc kubenswrapper[5109]: E1211 16:56:23.848064 5109 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.17:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18803798eb089159 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f7dbc7e1ee9c187a863ef9b473fad27b,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:68c07ee2fb6450c7b3b35bfdfc158dc475aaa0bcf9fba28b5e310d7e03355c04\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-11 16:56:23.846998361 +0000 UTC m=+221.526689867,LastTimestamp:2025-12-11 16:56:23.846998361 +0000 UTC m=+221.526689867,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 11 16:56:24 crc kubenswrapper[5109]: I1211 16:56:24.072452 5109 generic.go:358] "Generic (PLEG): container finished" podID="b4fc4ad8-69ad-4903-9936-b4936f5483af" containerID="bcc6951db5af30060d139dbaf039de6eeeedd437f251f8b24bbd118df45187e0" exitCode=0 Dec 11 16:56:24 crc kubenswrapper[5109]: I1211 16:56:24.072540 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"b4fc4ad8-69ad-4903-9936-b4936f5483af","Type":"ContainerDied","Data":"bcc6951db5af30060d139dbaf039de6eeeedd437f251f8b24bbd118df45187e0"} Dec 11 16:56:24 crc kubenswrapper[5109]: I1211 16:56:24.073323 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:24 crc kubenswrapper[5109]: I1211 16:56:24.074451 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f7dbc7e1ee9c187a863ef9b473fad27b","Type":"ContainerStarted","Data":"f30049d6c1c3c371a845ef3c263309a97e2adc045aa421f3b352e02c204d636a"} Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.090577 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.091302 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f7dbc7e1ee9c187a863ef9b473fad27b","Type":"ContainerStarted","Data":"54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e"} Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.092304 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:25 crc kubenswrapper[5109]: E1211 16:56:25.092430 5109 kubelet.go:3342] "Failed creating a mirror pod" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.499008 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.499553 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.533570 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4fc4ad8-69ad-4903-9936-b4936f5483af-kube-api-access\") pod \"b4fc4ad8-69ad-4903-9936-b4936f5483af\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.533612 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-var-lock\") pod \"b4fc4ad8-69ad-4903-9936-b4936f5483af\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.533768 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-kubelet-dir\") pod \"b4fc4ad8-69ad-4903-9936-b4936f5483af\" (UID: \"b4fc4ad8-69ad-4903-9936-b4936f5483af\") " Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.533819 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-var-lock" (OuterVolumeSpecName: "var-lock") pod "b4fc4ad8-69ad-4903-9936-b4936f5483af" (UID: "b4fc4ad8-69ad-4903-9936-b4936f5483af"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.533886 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b4fc4ad8-69ad-4903-9936-b4936f5483af" (UID: "b4fc4ad8-69ad-4903-9936-b4936f5483af"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.534200 5109 reconciler_common.go:299] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.534224 5109 reconciler_common.go:299] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/b4fc4ad8-69ad-4903-9936-b4936f5483af-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.540048 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4fc4ad8-69ad-4903-9936-b4936f5483af-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b4fc4ad8-69ad-4903-9936-b4936f5483af" (UID: "b4fc4ad8-69ad-4903-9936-b4936f5483af"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:56:25 crc kubenswrapper[5109]: I1211 16:56:25.635493 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b4fc4ad8-69ad-4903-9936-b4936f5483af-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:26 crc kubenswrapper[5109]: I1211 16:56:26.100590 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-12-crc" Dec 11 16:56:26 crc kubenswrapper[5109]: I1211 16:56:26.100588 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-12-crc" event={"ID":"b4fc4ad8-69ad-4903-9936-b4936f5483af","Type":"ContainerDied","Data":"068dcea8d8f6435901cd43b3699b58a87e17b89edc70c4446c11fbbf66b22238"} Dec 11 16:56:26 crc kubenswrapper[5109]: I1211 16:56:26.100680 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="068dcea8d8f6435901cd43b3699b58a87e17b89edc70c4446c11fbbf66b22238" Dec 11 16:56:26 crc kubenswrapper[5109]: I1211 16:56:26.100964 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:26 crc kubenswrapper[5109]: E1211 16:56:26.101680 5109 kubelet.go:3342] "Failed creating a mirror pod" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:56:26 crc kubenswrapper[5109]: I1211 16:56:26.126925 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.016506 5109 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.017631 5109 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.018076 5109 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.018344 5109 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.018797 5109 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:28 crc kubenswrapper[5109]: I1211 16:56:28.018909 5109 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.019536 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="200ms" Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.220580 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="400ms" Dec 11 16:56:28 crc kubenswrapper[5109]: I1211 16:56:28.541443 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" podUID="693595ff-e600-48ce-9d44-2f996cc27307" containerName="oauth-openshift" containerID="cri-o://e57f87ee47ac6a19a1505c54775ea70c7f2ef80da763216df55ef895048d5412" gracePeriod=15 Dec 11 16:56:28 crc kubenswrapper[5109]: E1211 16:56:28.621555 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="800ms" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.121878 5109 generic.go:358] "Generic (PLEG): container finished" podID="693595ff-e600-48ce-9d44-2f996cc27307" containerID="e57f87ee47ac6a19a1505c54775ea70c7f2ef80da763216df55ef895048d5412" exitCode=0 Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.122094 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" event={"ID":"693595ff-e600-48ce-9d44-2f996cc27307","Type":"ContainerDied","Data":"e57f87ee47ac6a19a1505c54775ea70c7f2ef80da763216df55ef895048d5412"} Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.122329 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" event={"ID":"693595ff-e600-48ce-9d44-2f996cc27307","Type":"ContainerDied","Data":"308e6159c7ca96fe7a06acf3961378c3d2ad623326b8d34b39037b73685f9a65"} Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.122343 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="308e6159c7ca96fe7a06acf3961378c3d2ad623326b8d34b39037b73685f9a65" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.147776 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.148395 5109 status_manager.go:895] "Failed to get status for pod" podUID="693595ff-e600-48ce-9d44-2f996cc27307" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-mr25z\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.148831 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285243 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-error\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285312 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-router-certs\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285504 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-audit-policies\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285592 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-login\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285632 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-trusted-ca-bundle\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285672 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-service-ca\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285808 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-ocp-branding-template\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285862 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-serving-cert\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.285989 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-session\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.286038 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-idp-0-file-data\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.286899 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-provider-selection\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.286997 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-cliconfig\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287043 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/693595ff-e600-48ce-9d44-2f996cc27307-audit-dir\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287067 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287109 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7xzzc\" (UniqueName: \"kubernetes.io/projected/693595ff-e600-48ce-9d44-2f996cc27307-kube-api-access-7xzzc\") pod \"693595ff-e600-48ce-9d44-2f996cc27307\" (UID: \"693595ff-e600-48ce-9d44-2f996cc27307\") " Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287446 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287468 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287493 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/693595ff-e600-48ce-9d44-2f996cc27307-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287587 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.287630 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.291765 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.292269 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.293970 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.294466 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.295590 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.296162 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/693595ff-e600-48ce-9d44-2f996cc27307-kube-api-access-7xzzc" (OuterVolumeSpecName: "kube-api-access-7xzzc") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "kube-api-access-7xzzc". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.296505 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.297346 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.298001 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "693595ff-e600-48ce-9d44-2f996cc27307" (UID: "693595ff-e600-48ce-9d44-2f996cc27307"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390204 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390299 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390329 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390357 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390386 5109 reconciler_common.go:299] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/693595ff-e600-48ce-9d44-2f996cc27307-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390411 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-7xzzc\" (UniqueName: \"kubernetes.io/projected/693595ff-e600-48ce-9d44-2f996cc27307-kube-api-access-7xzzc\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390434 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390462 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390487 5109 reconciler_common.go:299] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390597 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390624 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390653 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.390677 5109 reconciler_common.go:299] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/693595ff-e600-48ce-9d44-2f996cc27307-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 11 16:56:29 crc kubenswrapper[5109]: E1211 16:56:29.423250 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="1.6s" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.899918 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.901348 5109 status_manager.go:895] "Failed to get status for pod" podUID="693595ff-e600-48ce-9d44-2f996cc27307" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-mr25z\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.901940 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.916466 5109 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.916519 5109 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:29 crc kubenswrapper[5109]: E1211 16:56:29.917216 5109 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:29 crc kubenswrapper[5109]: I1211 16:56:29.917485 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:29 crc kubenswrapper[5109]: W1211 16:56:29.949243 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod57755cc5f99000cc11e193051474d4e2.slice/crio-56cbc097a41f6c79e372797a36188467488feb49fa4eeffdd586fc80da4a1ec1 WatchSource:0}: Error finding container 56cbc097a41f6c79e372797a36188467488feb49fa4eeffdd586fc80da4a1ec1: Status 404 returned error can't find the container with id 56cbc097a41f6c79e372797a36188467488feb49fa4eeffdd586fc80da4a1ec1 Dec 11 16:56:30 crc kubenswrapper[5109]: I1211 16:56:30.133777 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"56cbc097a41f6c79e372797a36188467488feb49fa4eeffdd586fc80da4a1ec1"} Dec 11 16:56:30 crc kubenswrapper[5109]: I1211 16:56:30.133896 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" Dec 11 16:56:30 crc kubenswrapper[5109]: I1211 16:56:30.136156 5109 status_manager.go:895] "Failed to get status for pod" podUID="693595ff-e600-48ce-9d44-2f996cc27307" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-mr25z\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:30 crc kubenswrapper[5109]: I1211 16:56:30.136670 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:30 crc kubenswrapper[5109]: I1211 16:56:30.168664 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:30 crc kubenswrapper[5109]: I1211 16:56:30.169029 5109 status_manager.go:895] "Failed to get status for pod" podUID="693595ff-e600-48ce-9d44-2f996cc27307" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-mr25z\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:31 crc kubenswrapper[5109]: E1211 16:56:31.024042 5109 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.17:6443: connect: connection refused" interval="3.2s" Dec 11 16:56:31 crc kubenswrapper[5109]: I1211 16:56:31.143982 5109 generic.go:358] "Generic (PLEG): container finished" podID="57755cc5f99000cc11e193051474d4e2" containerID="3317ad6e604d4b179ed6d9deb303b99e782285d8d890adaff2ec6023c40a3d6c" exitCode=0 Dec 11 16:56:31 crc kubenswrapper[5109]: I1211 16:56:31.144115 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerDied","Data":"3317ad6e604d4b179ed6d9deb303b99e782285d8d890adaff2ec6023c40a3d6c"} Dec 11 16:56:31 crc kubenswrapper[5109]: I1211 16:56:31.145420 5109 status_manager.go:895] "Failed to get status for pod" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" pod="openshift-kube-apiserver/installer-12-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-12-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:31 crc kubenswrapper[5109]: I1211 16:56:31.144529 5109 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:31 crc kubenswrapper[5109]: I1211 16:56:31.145669 5109 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:31 crc kubenswrapper[5109]: E1211 16:56:31.146196 5109 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.17:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:31 crc kubenswrapper[5109]: I1211 16:56:31.146255 5109 status_manager.go:895] "Failed to get status for pod" podUID="693595ff-e600-48ce-9d44-2f996cc27307" pod="openshift-authentication/oauth-openshift-66458b6674-mr25z" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-66458b6674-mr25z\": dial tcp 38.102.83.17:6443: connect: connection refused" Dec 11 16:56:32 crc kubenswrapper[5109]: I1211 16:56:32.156407 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"db6bb8836daa09532aed5867d37d21e85de17f6dd191080e8482161a6e06dea4"} Dec 11 16:56:32 crc kubenswrapper[5109]: I1211 16:56:32.156758 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"b4b294825e584c02ebc002589b2de0681a35eb9ce5520fd67b2de7ff8c1824d1"} Dec 11 16:56:32 crc kubenswrapper[5109]: I1211 16:56:32.156778 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"3fae0a3eccf1c63638687097780765ca75588fbf69b11316bea408a05dbaf7af"} Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.163176 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"d1c5dd44d69e215b228ec11f6b951a1336c8fb831a30d1b76d19740372ae749c"} Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.163392 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"57755cc5f99000cc11e193051474d4e2","Type":"ContainerStarted","Data":"bbf57cd500ad45dec5f3decf83eac176923ca826975b3306cb24b7f313dab023"} Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.163438 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.163566 5109 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.163590 5109 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.165985 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.166055 5109 generic.go:358] "Generic (PLEG): container finished" podID="9f0bc7fcb0822a2c13eb2d22cd8c0641" containerID="80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6" exitCode=1 Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.166142 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerDied","Data":"80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6"} Dec 11 16:56:33 crc kubenswrapper[5109]: I1211 16:56:33.166561 5109 scope.go:117] "RemoveContainer" containerID="80a2a5323569703d2eb0cbb35972dfeb77bf926e19890e367e6b1c39880344b6" Dec 11 16:56:34 crc kubenswrapper[5109]: I1211 16:56:34.173685 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 16:56:34 crc kubenswrapper[5109]: I1211 16:56:34.173846 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"9f0bc7fcb0822a2c13eb2d22cd8c0641","Type":"ContainerStarted","Data":"94695ab888b9756869fd24d88f968ae4e5cdc3c80275629be127844cf6481ce5"} Dec 11 16:56:34 crc kubenswrapper[5109]: I1211 16:56:34.917808 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:34 crc kubenswrapper[5109]: I1211 16:56:34.918176 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:34 crc kubenswrapper[5109]: I1211 16:56:34.928701 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:37 crc kubenswrapper[5109]: I1211 16:56:37.761880 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 16:56:37 crc kubenswrapper[5109]: I1211 16:56:37.762931 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 16:56:38 crc kubenswrapper[5109]: I1211 16:56:38.270347 5109 kubelet.go:3329] "Deleted mirror pod as it didn't match the static Pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:38 crc kubenswrapper[5109]: I1211 16:56:38.270395 5109 kubelet.go:3340] "Creating a mirror pod for static pod" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:38 crc kubenswrapper[5109]: I1211 16:56:38.313965 5109 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="eb32b105-b2ec-4ccc-a086-13fd5adfc098" Dec 11 16:56:39 crc kubenswrapper[5109]: I1211 16:56:39.210268 5109 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:39 crc kubenswrapper[5109]: I1211 16:56:39.210538 5109 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:39 crc kubenswrapper[5109]: I1211 16:56:39.213866 5109 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="eb32b105-b2ec-4ccc-a086-13fd5adfc098" Dec 11 16:56:39 crc kubenswrapper[5109]: I1211 16:56:39.217666 5109 status_manager.go:346] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://3fae0a3eccf1c63638687097780765ca75588fbf69b11316bea408a05dbaf7af" Dec 11 16:56:39 crc kubenswrapper[5109]: I1211 16:56:39.217735 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:39 crc kubenswrapper[5109]: I1211 16:56:39.475182 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:56:39 crc kubenswrapper[5109]: I1211 16:56:39.479628 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:56:40 crc kubenswrapper[5109]: I1211 16:56:40.218657 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:56:40 crc kubenswrapper[5109]: I1211 16:56:40.219059 5109 kubelet.go:3323] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:40 crc kubenswrapper[5109]: I1211 16:56:40.219089 5109 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4c5b2f1d-3171-42d0-8861-bdb91323d73b" Dec 11 16:56:40 crc kubenswrapper[5109]: I1211 16:56:40.223926 5109 status_manager.go:905] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="57755cc5f99000cc11e193051474d4e2" podUID="eb32b105-b2ec-4ccc-a086-13fd5adfc098" Dec 11 16:56:45 crc kubenswrapper[5109]: I1211 16:56:45.152238 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"serving-cert\"" Dec 11 16:56:48 crc kubenswrapper[5109]: I1211 16:56:48.219298 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"catalog-operator-serving-cert\"" Dec 11 16:56:48 crc kubenswrapper[5109]: I1211 16:56:48.584989 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-rbac-proxy\"" Dec 11 16:56:49 crc kubenswrapper[5109]: I1211 16:56:49.055422 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-dockercfg-4vdnc\"" Dec 11 16:56:49 crc kubenswrapper[5109]: I1211 16:56:49.661867 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"serving-cert\"" Dec 11 16:56:49 crc kubenswrapper[5109]: I1211 16:56:49.663342 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-node-metrics-cert\"" Dec 11 16:56:49 crc kubenswrapper[5109]: I1211 16:56:49.721327 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-config\"" Dec 11 16:56:49 crc kubenswrapper[5109]: I1211 16:56:49.793612 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-metrics\"" Dec 11 16:56:50 crc kubenswrapper[5109]: I1211 16:56:50.029606 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"audit-1\"" Dec 11 16:56:50 crc kubenswrapper[5109]: I1211 16:56:50.313567 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"default-dockercfg-9pgs7\"" Dec 11 16:56:50 crc kubenswrapper[5109]: I1211 16:56:50.625954 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-dockercfg-sw6nc\"" Dec 11 16:56:50 crc kubenswrapper[5109]: I1211 16:56:50.631689 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"kube-root-ca.crt\"" Dec 11 16:56:50 crc kubenswrapper[5109]: I1211 16:56:50.760877 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"package-server-manager-serving-cert\"" Dec 11 16:56:50 crc kubenswrapper[5109]: I1211 16:56:50.965722 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"cluster-image-registry-operator-dockercfg-ntnd7\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.000691 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.141369 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"trusted-ca-bundle\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.231291 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.277555 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-config\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.289792 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.316233 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-dockercfg-bf7fj\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.444372 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.651045 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"openshift-service-ca.crt\"" Dec 11 16:56:51 crc kubenswrapper[5109]: I1211 16:56:51.846076 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-console\"/\"networking-console-plugin\"" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.286747 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-node-dockercfg-l2v2m\"" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.448076 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"env-overrides\"" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.527467 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"openshift-service-ca.crt\"" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.674082 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"trusted-ca\"" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.893546 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-ca-bundle\"" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.911489 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"kube-root-ca.crt\"" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.912150 5109 reflector.go:430] "Caches populated" type="*v1.Service" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:56:52 crc kubenswrapper[5109]: I1211 16:56:52.957496 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-serving-cert\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.078570 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.091701 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.212784 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.382706 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.458255 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"openshift-service-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.540150 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.553704 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.619037 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"kube-root-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.621476 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-kubernetes-control-plane-dockercfg-nl8tp\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.669923 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-admission-controller-secret\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.677967 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\"" Dec 11 16:56:53 crc kubenswrapper[5109]: I1211 16:56:53.693318 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-wzhvk\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.067178 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-dockercfg-6c46w\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.188865 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"audit-1\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.231957 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-scheduler-operator-serving-cert\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.595275 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-jmhxf\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.697095 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"config-operator-serving-cert\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.699463 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-global-ca\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.709031 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.758103 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"serving-cert\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.808174 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"kube-root-ca.crt\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.918827 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-operator\"/\"trusted-ca\"" Dec 11 16:56:54 crc kubenswrapper[5109]: I1211 16:56:54.999448 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.003576 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-dockercfg-dzw6b\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.020834 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-tls\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.089830 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"default-dockercfg-hqpm5\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.154438 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"image-import-ca\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.166323 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"signing-key\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.189516 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"console-config\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.222430 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mcc-proxy-tls\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.278026 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"openshift-service-ca.crt\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.468268 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.579825 5109 reflector.go:430] "Caches populated" type="*v1.Node" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.581504 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-mmcpt\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.663000 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-route-controller-manager\"/\"serving-cert\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.782313 5109 reflector.go:430] "Caches populated" type="*v1.Pod" reflector="pkg/kubelet/config/apiserver.go:66" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.790628 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-66458b6674-mr25z","openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.790788 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.798676 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.814533 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=17.814511336 podStartE2EDuration="17.814511336s" podCreationTimestamp="2025-12-11 16:56:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:56:55.809067814 +0000 UTC m=+253.488759290" watchObservedRunningTime="2025-12-11 16:56:55.814511336 +0000 UTC m=+253.494202812" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.898258 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:55 crc kubenswrapper[5109]: I1211 16:56:55.907059 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"image-registry-operator-tls\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.023700 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"default-dockercfg-g6kgg\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.049708 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"openshift-service-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.086597 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ancillary-tools-dockercfg-nwglk\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.112863 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.114298 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"kube-root-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.142286 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"kube-root-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.226233 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serviceaccount-dockercfg-4gqzj\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.228577 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"kube-root-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.245946 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-default-metrics-tls\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.301898 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"installation-pull-secrets\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.403090 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-dockercfg-8dkm8\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.415700 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.496363 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.496494 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"serving-cert\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.525939 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"pprof-cert\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.552555 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.621213 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.632380 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"default-dockercfg-mdwwj\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.645436 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"client-ca\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.692512 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"kube-rbac-proxy\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.701619 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-operator-images\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.714952 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-console\"/\"networking-console-plugin-cert\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.721107 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-dockercfg-gnx66\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.741708 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"kube-root-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.770856 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"openshift-service-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.810125 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-apiserver-operator-config\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.838080 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-root-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.907233 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="693595ff-e600-48ce-9d44-2f996cc27307" path="/var/lib/kubelet/pods/693595ff-e600-48ce-9d44-2f996cc27307/volumes" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.910813 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"dns-operator-dockercfg-wbbsn\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.918783 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.951830 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"oauth-serving-cert\"" Dec 11 16:56:56 crc kubenswrapper[5109]: I1211 16:56:56.997906 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.047608 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"console-operator-dockercfg-kl6m8\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.075946 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-tls\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.083204 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.103356 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"signing-cabundle\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.130501 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.265304 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"env-overrides\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.366164 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"kube-root-ca.crt\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.529439 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-storage-version-migrator-operator-dockercfg-2h6bs\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.550818 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.683755 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-dockercfg-tnfx9\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.720964 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.786396 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"etcd-client\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.787686 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"kube-root-ca.crt\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.905769 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"multus-daemon-config\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.915853 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-version\"/\"openshift-service-ca.crt\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.925016 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"openshift-service-ca.crt\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.932872 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console\"/\"service-ca\"" Dec 11 16:56:57 crc kubenswrapper[5109]: I1211 16:56:57.998301 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress\"/\"service-ca-bundle\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.043231 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"kube-root-ca.crt\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.127904 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-client\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.146805 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"kube-root-ca.crt\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.165670 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"olm-operator-serving-cert\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.225232 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"dns-default\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.390685 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager\"/\"config\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.443059 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"hostpath-provisioner\"/\"csi-hostpath-provisioner-sa-dockercfg-7dcws\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.498601 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator\"/\"kube-storage-version-migrator-sa-dockercfg-kknhg\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.537040 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.561567 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"etcd-serving-ca\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.587595 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-daemon-dockercfg-w9nzh\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.666526 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"node-resolver-dockercfg-tk7bt\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.860134 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"config\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.938701 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-image-registry\"/\"image-registry-certificates\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.975447 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\"" Dec 11 16:56:58 crc kubenswrapper[5109]: I1211 16:56:58.978552 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.148410 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ovn-kubernetes\"/\"ovn-control-plane-metrics-cert\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.236173 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-config\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.241916 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca\"/\"openshift-service-ca.crt\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.266597 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"proxy-tls\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.313970 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-controller-dockercfg-xnj77\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.363493 5109 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.432554 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"console-operator-config\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.574319 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"machine-config-server-tls\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.595190 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"hostpath-provisioner\"/\"openshift-service-ca.crt\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.608168 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"kube-root-ca.crt\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.661993 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.669842 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.687811 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"default-cni-sysctl-allowlist\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.901242 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:56:59 crc kubenswrapper[5109]: I1211 16:56:59.926107 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-root-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.040353 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"kube-root-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.111794 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-djmfg\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.112938 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.128654 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-operator\"/\"metrics-tls\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.140872 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"node-ca-dockercfg-tjs74\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.150922 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns\"/\"dns-dockercfg-kpvmz\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.159013 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"iptables-alerter-script\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.178864 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"kube-root-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.214373 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-metrics-certs-default\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.221931 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns\"/\"openshift-service-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.246720 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"openshift-service-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.253854 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.422020 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-images\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.457663 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-service-ca-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.549886 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-service-ca-bundle\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.599536 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-config\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.648124 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-oauth-config\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.740258 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"kube-root-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.750221 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"config\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.755220 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-6n5ln\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.756209 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-4zqgh\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.810927 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-dockercfg-kw8fx\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.851693 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-dockercfg-jcmfj\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.856085 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-api\"/\"control-plane-machine-set-operator-tls\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.891317 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.905474 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"kube-root-ca.crt\"" Dec 11 16:57:00 crc kubenswrapper[5109]: I1211 16:57:00.939133 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver-operator\"/\"kube-root-ca.crt\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.022943 5109 kubelet.go:2547] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.023188 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" containerID="cri-o://54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e" gracePeriod=5 Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.026424 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.163872 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console-operator\"/\"serving-cert\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.226484 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"client-ca\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.300571 5109 reflector.go:430] "Caches populated" type="*v1.CSIDriver" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.313536 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-image-registry\"/\"registry-dockercfg-6w67b\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.356218 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"ingress-operator-dockercfg-74nwh\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.401869 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"openshift-service-ca.crt\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.467703 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.527412 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"trusted-ca-bundle\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.612181 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-sa-dockercfg-t8n29\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.742107 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca-operator\"/\"service-ca-operator-dockercfg-bjqfd\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.789898 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-console-operator\"/\"trusted-ca\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.879147 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-cluster-version\"/\"cluster-version-operator-serving-cert\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.880797 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"etcd-client\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.901986 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"metrics-daemon-secret\"" Dec 11 16:57:01 crc kubenswrapper[5109]: I1211 16:57:01.976858 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-certs-default\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.129688 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"config\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.159199 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ingress-canary\"/\"openshift-service-ca.crt\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.183106 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-multus\"/\"multus-ac-dockercfg-gj7jx\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.241849 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-dns-operator\"/\"metrics-tls\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.246770 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"packageserver-service-cert\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.338256 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-network-node-identity\"/\"network-node-identity-cert\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.399223 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-script-lib\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.400612 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"cni-copy-resources\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.418462 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress\"/\"router-stats-default\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.513525 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"marketplace-operator-dockercfg-2cfkp\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.598664 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-marketplace\"/\"marketplace-trusted-ca\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.631303 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"serving-cert\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.642652 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"serving-cert\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.800835 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-apiserver\"/\"encryption-config-1\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.894733 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-config\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.908523 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-d578988d4-qlfht"] Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909316 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909406 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909492 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" containerName="installer" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909571 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" containerName="installer" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909643 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="693595ff-e600-48ce-9d44-2f996cc27307" containerName="oauth-openshift" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909707 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="693595ff-e600-48ce-9d44-2f996cc27307" containerName="oauth-openshift" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909519 5109 status_manager.go:895] "Failed to get status for pod" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="pods \"kube-apiserver-startup-monitor-crc\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909924 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" containerName="startup-monitor" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.909984 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="b4fc4ad8-69ad-4903-9936-b4936f5483af" containerName="installer" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.910015 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="693595ff-e600-48ce-9d44-2f996cc27307" containerName="oauth-openshift" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.926238 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d578988d4-qlfht"] Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.926407 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.960282 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-controller-manager-operator-serving-cert\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.960476 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-login\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.960514 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-session\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.960497 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-idp-0-file-data\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.960568 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-provider-selection\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.962799 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"audit\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.964120 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-user-template-error\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.965540 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"oauth-openshift-dockercfg-d2bf2\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.966421 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-service-ca\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.966525 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"openshift-service-ca.crt\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.967220 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-serving-cert\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.967193 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"kube-root-ca.crt\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.967593 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-cliconfig\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.967858 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-router-certs\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.971547 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-diagnostics\"/\"kube-root-ca.crt\"" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978085 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-router-certs\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978194 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-error\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978261 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-session\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978315 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978364 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-audit-policies\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978520 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978642 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978700 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.978761 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9607b608-17f2-49a8-a0e3-80610a78b894-audit-dir\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.979033 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.979147 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.979351 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbc52\" (UniqueName: \"kubernetes.io/projected/9607b608-17f2-49a8-a0e3-80610a78b894-kube-api-access-jbc52\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.979646 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-login\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.979766 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-service-ca\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:02 crc kubenswrapper[5109]: I1211 16:57:02.998305 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-trusted-ca-bundle\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.004991 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.007298 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication\"/\"v4-0-config-system-ocp-branding-template\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.027564 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"trusted-ca-bundle\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.037653 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-config-operator\"/\"openshift-config-operator-dockercfg-sjn6s\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081010 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081092 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081157 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jbc52\" (UniqueName: \"kubernetes.io/projected/9607b608-17f2-49a8-a0e3-80610a78b894-kube-api-access-jbc52\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081205 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-login\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081241 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-service-ca\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081316 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-router-certs\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081493 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-error\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081535 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-session\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081566 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081597 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-audit-policies\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081641 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081694 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081733 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.081789 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9607b608-17f2-49a8-a0e3-80610a78b894-audit-dir\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.082643 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.082790 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-cliconfig\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.082865 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-service-ca\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.083630 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/9607b608-17f2-49a8-a0e3-80610a78b894-audit-policies\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.083709 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9607b608-17f2-49a8-a0e3-80610a78b894-audit-dir\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.101552 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-login\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.101603 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-error\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.101879 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-router-certs\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.101879 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-session\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.102035 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.102142 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.102263 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-system-serving-cert\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.102368 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/9607b608-17f2-49a8-a0e3-80610a78b894-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.108824 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbc52\" (UniqueName: \"kubernetes.io/projected/9607b608-17f2-49a8-a0e3-80610a78b894-kube-api-access-jbc52\") pod \"oauth-openshift-d578988d4-qlfht\" (UID: \"9607b608-17f2-49a8-a0e3-80610a78b894\") " pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.165205 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-config\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.310391 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.324899 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-config\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.505674 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-config-operator\"/\"kube-root-ca.crt\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.509690 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-apiserver\"/\"kube-root-ca.crt\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.602656 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-operator\"/\"metrics-tls\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.767126 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"openshift-service-ca.crt\"" Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.790996 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-d578988d4-qlfht"] Dec 11 16:57:03 crc kubenswrapper[5109]: I1211 16:57:03.853877 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-node-identity\"/\"ovnkube-identity-cm\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.029093 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-controller-manager-operator\"/\"kube-controller-manager-operator-serving-cert\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.090802 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-authentication-operator\"/\"service-ca-bundle\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.246489 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-dns-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.277978 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.308236 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-ingress-canary\"/\"canary-serving-cert\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.388850 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-network-operator\"/\"kube-root-ca.crt\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.395382 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.395458 5109 generic.go:358] "Generic (PLEG): container finished" podID="9607b608-17f2-49a8-a0e3-80610a78b894" containerID="b5b7747fedcfc167ef3719e519bc7eb50e6378d834a4451341bf77e5e86bf454" exitCode=255 Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.395661 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" event={"ID":"9607b608-17f2-49a8-a0e3-80610a78b894","Type":"ContainerDied","Data":"b5b7747fedcfc167ef3719e519bc7eb50e6378d834a4451341bf77e5e86bf454"} Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.395733 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" event={"ID":"9607b608-17f2-49a8-a0e3-80610a78b894","Type":"ContainerStarted","Data":"c230172e866f74d27043e9f1973c32a1179a961489a487e7df6abee9b0ec2a71"} Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.396510 5109 scope.go:117] "RemoveContainer" containerID="b5b7747fedcfc167ef3719e519bc7eb50e6378d834a4451341bf77e5e86bf454" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.684146 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-service-ca\"/\"service-ca-dockercfg-bgxvm\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.709431 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-console\"/\"console-serving-cert\"" Dec 11 16:57:04 crc kubenswrapper[5109]: I1211 16:57:04.959057 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-multus\"/\"whereabouts-flatfile-config\"" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.025485 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-kube-storage-version-migrator-operator\"/\"kube-root-ca.crt\"" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.043191 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"node-bootstrapper-token\"" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.148313 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-apiserver-operator\"/\"kube-apiserver-operator-serving-cert\"" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.336765 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-ovn-kubernetes\"/\"ovnkube-config\"" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.403759 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.403917 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" event={"ID":"9607b608-17f2-49a8-a0e3-80610a78b894","Type":"ContainerStarted","Data":"9808e1a87d99740d08b39442e65fefb518a24a06895449bf6f3d483a24f269d6"} Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.404413 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.414035 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.430118 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-d578988d4-qlfht" podStartSLOduration=62.43010104 podStartE2EDuration="1m2.43010104s" podCreationTimestamp="2025-12-11 16:56:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:57:05.429055391 +0000 UTC m=+263.108746867" watchObservedRunningTime="2025-12-11 16:57:05.43010104 +0000 UTC m=+263.109792516" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.477844 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-controller-manager-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.520405 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-machine-config-operator\"/\"openshift-service-ca.crt\"" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.801978 5109 reflector.go:430] "Caches populated" type="*v1.RuntimeClass" reflector="k8s.io/client-go/informers/factory.go:160" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.923947 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55806: no serving certificate available for the kubelet" Dec 11 16:57:05 crc kubenswrapper[5109]: I1211 16:57:05.975783 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-machine-config-operator\"/\"mco-proxy-tls\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.076811 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-etcd-operator\"/\"etcd-operator-serving-cert\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.138298 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.138375 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.139603 5109 status_manager.go:895] "Failed to get status for pod" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="pods \"kube-apiserver-startup-monitor-crc\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.233892 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234026 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234024 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock" (OuterVolumeSpecName: "var-lock") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234088 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234121 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log" (OuterVolumeSpecName: "var-log") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234126 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234185 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234227 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests" (OuterVolumeSpecName: "manifests") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234292 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") pod \"f7dbc7e1ee9c187a863ef9b473fad27b\" (UID: \"f7dbc7e1ee9c187a863ef9b473fad27b\") " Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234906 5109 reconciler_common.go:299] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-log\") on node \"crc\" DevicePath \"\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234943 5109 reconciler_common.go:299] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-manifests\") on node \"crc\" DevicePath \"\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234967 5109 reconciler_common.go:299] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.234989 5109 reconciler_common.go:299] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-var-lock\") on node \"crc\" DevicePath \"\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.244019 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f7dbc7e1ee9c187a863ef9b473fad27b" (UID: "f7dbc7e1ee9c187a863ef9b473fad27b"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.336132 5109 reconciler_common.go:299] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f7dbc7e1ee9c187a863ef9b473fad27b-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.416585 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f7dbc7e1ee9c187a863ef9b473fad27b/startup-monitor/0.log" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.416653 5109 generic.go:358] "Generic (PLEG): container finished" podID="f7dbc7e1ee9c187a863ef9b473fad27b" containerID="54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e" exitCode=137 Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.417068 5109 scope.go:117] "RemoveContainer" containerID="54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.417452 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.432384 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-authentication-operator\"/\"authentication-operator-dockercfg-6tbpn\"" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.452034 5109 scope.go:117] "RemoveContainer" containerID="54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e" Dec 11 16:57:06 crc kubenswrapper[5109]: E1211 16:57:06.453442 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e\": container with ID starting with 54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e not found: ID does not exist" containerID="54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.453531 5109 status_manager.go:895] "Failed to get status for pod" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="pods \"kube-apiserver-startup-monitor-crc\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.453525 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e"} err="failed to get container status \"54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e\": rpc error: code = NotFound desc = could not find container \"54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e\": container with ID starting with 54e5229270a49867cd3316fe71ce6570d8a9b5ba278a0ebd47150138a4762f0e not found: ID does not exist" Dec 11 16:57:06 crc kubenswrapper[5109]: I1211 16:57:06.913100 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7dbc7e1ee9c187a863ef9b473fad27b" path="/var/lib/kubelet/pods/f7dbc7e1ee9c187a863ef9b473fad27b/volumes" Dec 11 16:57:07 crc kubenswrapper[5109]: I1211 16:57:07.601016 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-qqw4z\"" Dec 11 16:57:07 crc kubenswrapper[5109]: I1211 16:57:07.729758 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-kube-scheduler-operator\"/\"openshift-kube-scheduler-operator-dockercfg-2wbn2\"" Dec 11 16:57:07 crc kubenswrapper[5109]: I1211 16:57:07.762919 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 16:57:07 crc kubenswrapper[5109]: I1211 16:57:07.762993 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 16:57:22 crc kubenswrapper[5109]: I1211 16:57:22.369515 5109 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-tzg7t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" start-of-body= Dec 11 16:57:22 crc kubenswrapper[5109]: I1211 16:57:22.370246 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": dial tcp 10.217.0.41:8080: connect: connection refused" Dec 11 16:57:22 crc kubenswrapper[5109]: I1211 16:57:22.533547 5109 generic.go:358] "Generic (PLEG): container finished" podID="28a5c354-8565-49aa-b329-cf529c594431" containerID="2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849" exitCode=0 Dec 11 16:57:22 crc kubenswrapper[5109]: I1211 16:57:22.533673 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" event={"ID":"28a5c354-8565-49aa-b329-cf529c594431","Type":"ContainerDied","Data":"2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849"} Dec 11 16:57:22 crc kubenswrapper[5109]: I1211 16:57:22.534373 5109 scope.go:117] "RemoveContainer" containerID="2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849" Dec 11 16:57:23 crc kubenswrapper[5109]: I1211 16:57:23.544978 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" event={"ID":"28a5c354-8565-49aa-b329-cf529c594431","Type":"ContainerStarted","Data":"e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d"} Dec 11 16:57:23 crc kubenswrapper[5109]: I1211 16:57:23.546380 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:57:23 crc kubenswrapper[5109]: I1211 16:57:23.549160 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:57:34 crc kubenswrapper[5109]: I1211 16:57:34.581648 5109 ???:1] "http: TLS handshake error from 192.168.126.11:44302: no serving certificate available for the kubelet" Dec 11 16:57:37 crc kubenswrapper[5109]: I1211 16:57:37.762794 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 16:57:37 crc kubenswrapper[5109]: I1211 16:57:37.763162 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 16:57:37 crc kubenswrapper[5109]: I1211 16:57:37.763229 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 16:57:37 crc kubenswrapper[5109]: I1211 16:57:37.764046 5109 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f099a425a82983d885b200b73fcf89b1dba37bc85eba89c7f84a71efcc7178c7"} pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 16:57:37 crc kubenswrapper[5109]: I1211 16:57:37.764147 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" containerID="cri-o://f099a425a82983d885b200b73fcf89b1dba37bc85eba89c7f84a71efcc7178c7" gracePeriod=600 Dec 11 16:57:38 crc kubenswrapper[5109]: I1211 16:57:38.662467 5109 generic.go:358] "Generic (PLEG): container finished" podID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerID="f099a425a82983d885b200b73fcf89b1dba37bc85eba89c7f84a71efcc7178c7" exitCode=0 Dec 11 16:57:38 crc kubenswrapper[5109]: I1211 16:57:38.662633 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerDied","Data":"f099a425a82983d885b200b73fcf89b1dba37bc85eba89c7f84a71efcc7178c7"} Dec 11 16:57:39 crc kubenswrapper[5109]: I1211 16:57:39.673913 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"dfe59f5af7913b6af69c36c59fa1989b46d095e49ddef2e9386892be6189ace2"} Dec 11 16:57:43 crc kubenswrapper[5109]: I1211 16:57:43.029274 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 16:57:43 crc kubenswrapper[5109]: I1211 16:57:43.029985 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 16:57:43 crc kubenswrapper[5109]: I1211 16:57:43.111295 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 16:57:43 crc kubenswrapper[5109]: I1211 16:57:43.111643 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 16:58:17 crc kubenswrapper[5109]: I1211 16:58:17.761177 5109 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.038723 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7fhv"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.039900 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t7fhv" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="registry-server" containerID="cri-o://f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d" gracePeriod=30 Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.044668 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6ctk"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.044945 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v6ctk" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="registry-server" containerID="cri-o://92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375" gracePeriod=30 Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.056590 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-tzg7t"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.057278 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" containerID="cri-o://e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d" gracePeriod=30 Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.063333 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txfkl"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.063618 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-txfkl" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="registry-server" containerID="cri-o://947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c" gracePeriod=30 Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.074186 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w9lmv"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.074625 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w9lmv" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="registry-server" containerID="cri-o://d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3" gracePeriod=30 Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.083394 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-hnv7m"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.091693 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.098489 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-hnv7m"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.147587 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/968a8805-21c9-4337-8a07-f85ee689a248-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.147634 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmgns\" (UniqueName: \"kubernetes.io/projected/968a8805-21c9-4337-8a07-f85ee689a248-kube-api-access-wmgns\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.147786 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/968a8805-21c9-4337-8a07-f85ee689a248-tmp\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.147831 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/968a8805-21c9-4337-8a07-f85ee689a248-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.249379 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/968a8805-21c9-4337-8a07-f85ee689a248-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.249454 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wmgns\" (UniqueName: \"kubernetes.io/projected/968a8805-21c9-4337-8a07-f85ee689a248-kube-api-access-wmgns\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.251260 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/968a8805-21c9-4337-8a07-f85ee689a248-tmp\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.251286 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/968a8805-21c9-4337-8a07-f85ee689a248-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.251653 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/968a8805-21c9-4337-8a07-f85ee689a248-tmp\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.252942 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/968a8805-21c9-4337-8a07-f85ee689a248-marketplace-trusted-ca\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.281477 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/968a8805-21c9-4337-8a07-f85ee689a248-marketplace-operator-metrics\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.283646 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmgns\" (UniqueName: \"kubernetes.io/projected/968a8805-21c9-4337-8a07-f85ee689a248-kube-api-access-wmgns\") pod \"marketplace-operator-547dbd544d-hnv7m\" (UID: \"968a8805-21c9-4337-8a07-f85ee689a248\") " pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.444819 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.452963 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.519900 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.534531 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.556408 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.565094 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568402 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-utilities\") pod \"591b7ef7-10f4-43e0-9927-bb5ce699351c\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568436 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-catalog-content\") pod \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568502 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-utilities\") pod \"d76f4925-c474-4764-9e50-1597f45a32b6\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568526 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-catalog-content\") pod \"591b7ef7-10f4-43e0-9927-bb5ce699351c\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568544 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kl4v4\" (UniqueName: \"kubernetes.io/projected/591b7ef7-10f4-43e0-9927-bb5ce699351c-kube-api-access-kl4v4\") pod \"591b7ef7-10f4-43e0-9927-bb5ce699351c\" (UID: \"591b7ef7-10f4-43e0-9927-bb5ce699351c\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568576 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-utilities\") pod \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568600 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jfxwh\" (UniqueName: \"kubernetes.io/projected/d76f4925-c474-4764-9e50-1597f45a32b6-kube-api-access-jfxwh\") pod \"d76f4925-c474-4764-9e50-1597f45a32b6\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568651 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qp65\" (UniqueName: \"kubernetes.io/projected/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-kube-api-access-4qp65\") pod \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\" (UID: \"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.568682 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-catalog-content\") pod \"d76f4925-c474-4764-9e50-1597f45a32b6\" (UID: \"d76f4925-c474-4764-9e50-1597f45a32b6\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.570672 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-utilities" (OuterVolumeSpecName: "utilities") pod "591b7ef7-10f4-43e0-9927-bb5ce699351c" (UID: "591b7ef7-10f4-43e0-9927-bb5ce699351c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.573456 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-utilities" (OuterVolumeSpecName: "utilities") pod "d76f4925-c474-4764-9e50-1597f45a32b6" (UID: "d76f4925-c474-4764-9e50-1597f45a32b6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.573501 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-utilities" (OuterVolumeSpecName: "utilities") pod "38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" (UID: "38d2b42b-3ce1-4f8b-994d-c9a69e5f8363"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.577442 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d76f4925-c474-4764-9e50-1597f45a32b6-kube-api-access-jfxwh" (OuterVolumeSpecName: "kube-api-access-jfxwh") pod "d76f4925-c474-4764-9e50-1597f45a32b6" (UID: "d76f4925-c474-4764-9e50-1597f45a32b6"). InnerVolumeSpecName "kube-api-access-jfxwh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.581914 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-kube-api-access-4qp65" (OuterVolumeSpecName: "kube-api-access-4qp65") pod "38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" (UID: "38d2b42b-3ce1-4f8b-994d-c9a69e5f8363"). InnerVolumeSpecName "kube-api-access-4qp65". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.582015 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/591b7ef7-10f4-43e0-9927-bb5ce699351c-kube-api-access-kl4v4" (OuterVolumeSpecName: "kube-api-access-kl4v4") pod "591b7ef7-10f4-43e0-9927-bb5ce699351c" (UID: "591b7ef7-10f4-43e0-9927-bb5ce699351c"). InnerVolumeSpecName "kube-api-access-kl4v4". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.619092 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" (UID: "38d2b42b-3ce1-4f8b-994d-c9a69e5f8363"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.626172 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "591b7ef7-10f4-43e0-9927-bb5ce699351c" (UID: "591b7ef7-10f4-43e0-9927-bb5ce699351c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669329 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdxv9\" (UniqueName: \"kubernetes.io/projected/12530ca5-956e-43fd-9b42-40f50a32c8b8-kube-api-access-zdxv9\") pod \"12530ca5-956e-43fd-9b42-40f50a32c8b8\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669375 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zv59g\" (UniqueName: \"kubernetes.io/projected/28a5c354-8565-49aa-b329-cf529c594431-kube-api-access-zv59g\") pod \"28a5c354-8565-49aa-b329-cf529c594431\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669419 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/28a5c354-8565-49aa-b329-cf529c594431-marketplace-operator-metrics\") pod \"28a5c354-8565-49aa-b329-cf529c594431\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669460 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca\") pod \"28a5c354-8565-49aa-b329-cf529c594431\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669505 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-utilities\") pod \"12530ca5-956e-43fd-9b42-40f50a32c8b8\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669551 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-catalog-content\") pod \"12530ca5-956e-43fd-9b42-40f50a32c8b8\" (UID: \"12530ca5-956e-43fd-9b42-40f50a32c8b8\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669585 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/28a5c354-8565-49aa-b329-cf529c594431-tmp\") pod \"28a5c354-8565-49aa-b329-cf529c594431\" (UID: \"28a5c354-8565-49aa-b329-cf529c594431\") " Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669772 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669787 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669796 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669804 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/591b7ef7-10f4-43e0-9927-bb5ce699351c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669812 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kl4v4\" (UniqueName: \"kubernetes.io/projected/591b7ef7-10f4-43e0-9927-bb5ce699351c-kube-api-access-kl4v4\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669822 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669830 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-jfxwh\" (UniqueName: \"kubernetes.io/projected/d76f4925-c474-4764-9e50-1597f45a32b6-kube-api-access-jfxwh\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.669839 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-4qp65\" (UniqueName: \"kubernetes.io/projected/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363-kube-api-access-4qp65\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.671710 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28a5c354-8565-49aa-b329-cf529c594431-tmp" (OuterVolumeSpecName: "tmp") pod "28a5c354-8565-49aa-b329-cf529c594431" (UID: "28a5c354-8565-49aa-b329-cf529c594431"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.671991 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "28a5c354-8565-49aa-b329-cf529c594431" (UID: "28a5c354-8565-49aa-b329-cf529c594431"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.673894 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28a5c354-8565-49aa-b329-cf529c594431-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "28a5c354-8565-49aa-b329-cf529c594431" (UID: "28a5c354-8565-49aa-b329-cf529c594431"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.674637 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28a5c354-8565-49aa-b329-cf529c594431-kube-api-access-zv59g" (OuterVolumeSpecName: "kube-api-access-zv59g") pod "28a5c354-8565-49aa-b329-cf529c594431" (UID: "28a5c354-8565-49aa-b329-cf529c594431"). InnerVolumeSpecName "kube-api-access-zv59g". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.674687 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12530ca5-956e-43fd-9b42-40f50a32c8b8-kube-api-access-zdxv9" (OuterVolumeSpecName: "kube-api-access-zdxv9") pod "12530ca5-956e-43fd-9b42-40f50a32c8b8" (UID: "12530ca5-956e-43fd-9b42-40f50a32c8b8"). InnerVolumeSpecName "kube-api-access-zdxv9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.681545 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-utilities" (OuterVolumeSpecName: "utilities") pod "12530ca5-956e-43fd-9b42-40f50a32c8b8" (UID: "12530ca5-956e-43fd-9b42-40f50a32c8b8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.687300 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d76f4925-c474-4764-9e50-1597f45a32b6" (UID: "d76f4925-c474-4764-9e50-1597f45a32b6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.744887 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12530ca5-956e-43fd-9b42-40f50a32c8b8" (UID: "12530ca5-956e-43fd-9b42-40f50a32c8b8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770600 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770626 5109 reconciler_common.go:299] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/28a5c354-8565-49aa-b329-cf529c594431-tmp\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770636 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zdxv9\" (UniqueName: \"kubernetes.io/projected/12530ca5-956e-43fd-9b42-40f50a32c8b8-kube-api-access-zdxv9\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770646 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zv59g\" (UniqueName: \"kubernetes.io/projected/28a5c354-8565-49aa-b329-cf529c594431-kube-api-access-zv59g\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770656 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d76f4925-c474-4764-9e50-1597f45a32b6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770666 5109 reconciler_common.go:299] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/28a5c354-8565-49aa-b329-cf529c594431-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770674 5109 reconciler_common.go:299] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28a5c354-8565-49aa-b329-cf529c594431-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.770682 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12530ca5-956e-43fd-9b42-40f50a32c8b8-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.907845 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-hnv7m"] Dec 11 16:58:33 crc kubenswrapper[5109]: I1211 16:58:33.910899 5109 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.064140 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" event={"ID":"968a8805-21c9-4337-8a07-f85ee689a248","Type":"ContainerStarted","Data":"e14677d4d9ad5453f97da4dac942081c467ef2532375f25a5035ea7b74dfa06a"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.064181 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" event={"ID":"968a8805-21c9-4337-8a07-f85ee689a248","Type":"ContainerStarted","Data":"5106692953c53d59d6c9713554df311869dd187fcf1b69e79d851a130ed9f119"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.065292 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.066753 5109 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-hnv7m container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.67:8080/healthz\": dial tcp 10.217.0.67:8080: connect: connection refused" start-of-body= Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.066794 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" podUID="968a8805-21c9-4337-8a07-f85ee689a248" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.67:8080/healthz\": dial tcp 10.217.0.67:8080: connect: connection refused" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.068308 5109 generic.go:358] "Generic (PLEG): container finished" podID="28a5c354-8565-49aa-b329-cf529c594431" containerID="e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d" exitCode=0 Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.068352 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.068366 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" event={"ID":"28a5c354-8565-49aa-b329-cf529c594431","Type":"ContainerDied","Data":"e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.068671 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" event={"ID":"28a5c354-8565-49aa-b329-cf529c594431","Type":"ContainerDied","Data":"4f7462522b8a42127e1635bd5700c1e24e9c3cd31ffb9d71a4a78d23501d402d"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.068709 5109 scope.go:117] "RemoveContainer" containerID="e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.070131 5109 generic.go:358] "Generic (PLEG): container finished" podID="d76f4925-c474-4764-9e50-1597f45a32b6" containerID="d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3" exitCode=0 Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.070332 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w9lmv" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.070351 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9lmv" event={"ID":"d76f4925-c474-4764-9e50-1597f45a32b6","Type":"ContainerDied","Data":"d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.070420 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w9lmv" event={"ID":"d76f4925-c474-4764-9e50-1597f45a32b6","Type":"ContainerDied","Data":"a78cce0154ad46366334460dc0db4ca0c4dfb803dfe6ff1f22b178dfb6b877a4"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.072564 5109 generic.go:358] "Generic (PLEG): container finished" podID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerID="f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d" exitCode=0 Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.072679 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7fhv" event={"ID":"591b7ef7-10f4-43e0-9927-bb5ce699351c","Type":"ContainerDied","Data":"f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.072706 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t7fhv" event={"ID":"591b7ef7-10f4-43e0-9927-bb5ce699351c","Type":"ContainerDied","Data":"f8eb8d7af2768f7390b51e5f4f4aefa9d0c08da6e3b1ae37a434dce748de2390"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.072799 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t7fhv" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.076983 5109 generic.go:358] "Generic (PLEG): container finished" podID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerID="947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c" exitCode=0 Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.077014 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txfkl" event={"ID":"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363","Type":"ContainerDied","Data":"947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.077047 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txfkl" event={"ID":"38d2b42b-3ce1-4f8b-994d-c9a69e5f8363","Type":"ContainerDied","Data":"e104531d7ff56beae3ee5ffa0e3210ca03b83fc8b61dd3bbad955223b3f237f2"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.077074 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txfkl" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.078522 5109 generic.go:358] "Generic (PLEG): container finished" podID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerID="92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375" exitCode=0 Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.078583 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerDied","Data":"92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.078601 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v6ctk" event={"ID":"12530ca5-956e-43fd-9b42-40f50a32c8b8","Type":"ContainerDied","Data":"7ccd45c7567db516dd1ed7083f3dbfb1f31f8e54fceecbbc4780edfd61289617"} Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.078706 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v6ctk" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.083353 5109 scope.go:117] "RemoveContainer" containerID="2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.089995 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" podStartSLOduration=1.089973299 podStartE2EDuration="1.089973299s" podCreationTimestamp="2025-12-11 16:58:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 16:58:34.084869731 +0000 UTC m=+351.764561197" watchObservedRunningTime="2025-12-11 16:58:34.089973299 +0000 UTC m=+351.769664765" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.109077 5109 scope.go:117] "RemoveContainer" containerID="e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.109362 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d\": container with ID starting with e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d not found: ID does not exist" containerID="e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.109388 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d"} err="failed to get container status \"e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d\": rpc error: code = NotFound desc = could not find container \"e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d\": container with ID starting with e571add64ada66942eaac60d6b6a0cf02d702c01083d45907f398e48ed10f38d not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.109405 5109 scope.go:117] "RemoveContainer" containerID="2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.109655 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849\": container with ID starting with 2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849 not found: ID does not exist" containerID="2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.110288 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849"} err="failed to get container status \"2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849\": rpc error: code = NotFound desc = could not find container \"2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849\": container with ID starting with 2f5f244180e786f3a6390846299046676418f8099de6a0098206b5d31f1ba849 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.110332 5109 scope.go:117] "RemoveContainer" containerID="d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.123031 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-tzg7t"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.134039 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-547dbd544d-tzg7t"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.143835 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v6ctk"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.145879 5109 scope.go:117] "RemoveContainer" containerID="a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.156890 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v6ctk"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.163664 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t7fhv"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.167010 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t7fhv"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.168942 5109 scope.go:117] "RemoveContainer" containerID="5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.170341 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txfkl"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.173248 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-txfkl"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.176176 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w9lmv"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.179101 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w9lmv"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.188395 5109 scope.go:117] "RemoveContainer" containerID="d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.188785 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3\": container with ID starting with d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3 not found: ID does not exist" containerID="d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.188878 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3"} err="failed to get container status \"d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3\": rpc error: code = NotFound desc = could not find container \"d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3\": container with ID starting with d4785e32f2871ac6458d9cd9b18f0d6084a6a71fea57a9c8c274377eecce1be3 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.188963 5109 scope.go:117] "RemoveContainer" containerID="a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.189303 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b\": container with ID starting with a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b not found: ID does not exist" containerID="a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.189374 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b"} err="failed to get container status \"a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b\": rpc error: code = NotFound desc = could not find container \"a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b\": container with ID starting with a4c3bf8215242fd0c560fb9e5d62df13d69fe40cf2dcd7750d64e79091a7c48b not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.189441 5109 scope.go:117] "RemoveContainer" containerID="5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.189863 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4\": container with ID starting with 5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4 not found: ID does not exist" containerID="5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.189882 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4"} err="failed to get container status \"5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4\": rpc error: code = NotFound desc = could not find container \"5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4\": container with ID starting with 5c5cf1bf323c4dad8fe50cb156280ca70b451d7c7996b4e416c268e6942c77f4 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.189895 5109 scope.go:117] "RemoveContainer" containerID="f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.205170 5109 scope.go:117] "RemoveContainer" containerID="b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.228994 5109 scope.go:117] "RemoveContainer" containerID="e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.271364 5109 scope.go:117] "RemoveContainer" containerID="f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.271703 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d\": container with ID starting with f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d not found: ID does not exist" containerID="f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.271852 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d"} err="failed to get container status \"f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d\": rpc error: code = NotFound desc = could not find container \"f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d\": container with ID starting with f566722d80f704107dba216a737ddcc4e662fa81da99c3c8627810a78167756d not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.272056 5109 scope.go:117] "RemoveContainer" containerID="b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.272668 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6\": container with ID starting with b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6 not found: ID does not exist" containerID="b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.272691 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6"} err="failed to get container status \"b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6\": rpc error: code = NotFound desc = could not find container \"b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6\": container with ID starting with b6f6086b82fc65ecaf192eb8fa749f8fb1a3dfb7afcf1a7ecca3dec9ae3831a6 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.272706 5109 scope.go:117] "RemoveContainer" containerID="e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.273044 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621\": container with ID starting with e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621 not found: ID does not exist" containerID="e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.273087 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621"} err="failed to get container status \"e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621\": rpc error: code = NotFound desc = could not find container \"e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621\": container with ID starting with e3254fa41fe5523f9b520fdf72f9f6314f45e1fc064af3bff5b6257ea877a621 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.273115 5109 scope.go:117] "RemoveContainer" containerID="947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.287923 5109 scope.go:117] "RemoveContainer" containerID="7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.309788 5109 scope.go:117] "RemoveContainer" containerID="1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.337399 5109 scope.go:117] "RemoveContainer" containerID="947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.337951 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c\": container with ID starting with 947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c not found: ID does not exist" containerID="947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.337977 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c"} err="failed to get container status \"947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c\": rpc error: code = NotFound desc = could not find container \"947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c\": container with ID starting with 947ad26f154c6998f3fae89e2a76d14719c53deb77552223db9d080733d19c9c not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.337997 5109 scope.go:117] "RemoveContainer" containerID="7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.338464 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5\": container with ID starting with 7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5 not found: ID does not exist" containerID="7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.338488 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5"} err="failed to get container status \"7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5\": rpc error: code = NotFound desc = could not find container \"7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5\": container with ID starting with 7f1164dfb00e87f37a3b3265c86b8a28b8d6732547de6ce6d9156ba93d7691e5 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.338502 5109 scope.go:117] "RemoveContainer" containerID="1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.338893 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a\": container with ID starting with 1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a not found: ID does not exist" containerID="1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.338937 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a"} err="failed to get container status \"1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a\": rpc error: code = NotFound desc = could not find container \"1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a\": container with ID starting with 1d2e6264cb4e5198f02d418b156cc118cabe922bb5d65d621a2995769fafb28a not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.338962 5109 scope.go:117] "RemoveContainer" containerID="92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.355320 5109 scope.go:117] "RemoveContainer" containerID="0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.368605 5109 scope.go:117] "RemoveContainer" containerID="b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.383774 5109 scope.go:117] "RemoveContainer" containerID="92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.384301 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375\": container with ID starting with 92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375 not found: ID does not exist" containerID="92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.384338 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375"} err="failed to get container status \"92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375\": rpc error: code = NotFound desc = could not find container \"92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375\": container with ID starting with 92eb3b8885745137af20fd54c0421b97d917cdec86c24eacb989d32af4830375 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.384360 5109 scope.go:117] "RemoveContainer" containerID="0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.384691 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12\": container with ID starting with 0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12 not found: ID does not exist" containerID="0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.384711 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12"} err="failed to get container status \"0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12\": rpc error: code = NotFound desc = could not find container \"0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12\": container with ID starting with 0020e5af03c68840e43caa7065477d22f052adf57b6f0bf3f09532ad9c1e4e12 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.384724 5109 scope.go:117] "RemoveContainer" containerID="b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85" Dec 11 16:58:34 crc kubenswrapper[5109]: E1211 16:58:34.384983 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85\": container with ID starting with b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85 not found: ID does not exist" containerID="b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.385024 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85"} err="failed to get container status \"b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85\": rpc error: code = NotFound desc = could not find container \"b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85\": container with ID starting with b0903c6bf984bbc488c81e32e43a11dbda997eca349ff874a6131beb61e41c85 not found: ID does not exist" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.550657 5109 patch_prober.go:28] interesting pod/marketplace-operator-547dbd544d-tzg7t container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.41:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.550788 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-547dbd544d-tzg7t" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.41:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.855171 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-czz8n"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.856230 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.856256 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.856270 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.856307 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857360 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857382 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857432 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857442 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857454 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857463 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857474 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857511 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857527 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857533 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857544 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857551 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857588 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857597 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="extract-content" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857606 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857612 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857622 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857630 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857643 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857702 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857714 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857777 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="extract-utilities" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.857984 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.858022 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.858032 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.858043 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.858051 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.858109 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" containerName="registry-server" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.858279 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.858290 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="28a5c354-8565-49aa-b329-cf529c594431" containerName="marketplace-operator" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.871179 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz8n"] Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.871347 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.873665 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-marketplace-dockercfg-gg4w7\"" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.907038 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12530ca5-956e-43fd-9b42-40f50a32c8b8" path="/var/lib/kubelet/pods/12530ca5-956e-43fd-9b42-40f50a32c8b8/volumes" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.907822 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28a5c354-8565-49aa-b329-cf529c594431" path="/var/lib/kubelet/pods/28a5c354-8565-49aa-b329-cf529c594431/volumes" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.908246 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38d2b42b-3ce1-4f8b-994d-c9a69e5f8363" path="/var/lib/kubelet/pods/38d2b42b-3ce1-4f8b-994d-c9a69e5f8363/volumes" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.909174 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="591b7ef7-10f4-43e0-9927-bb5ce699351c" path="/var/lib/kubelet/pods/591b7ef7-10f4-43e0-9927-bb5ce699351c/volumes" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.909696 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d76f4925-c474-4764-9e50-1597f45a32b6" path="/var/lib/kubelet/pods/d76f4925-c474-4764-9e50-1597f45a32b6/volumes" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.984384 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-utilities\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.984622 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xw99k\" (UniqueName: \"kubernetes.io/projected/e1f282da-749c-42a0-aea4-0ec0587bc5fc-kube-api-access-xw99k\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:34 crc kubenswrapper[5109]: I1211 16:58:34.984865 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-catalog-content\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.085770 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-catalog-content\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.085882 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-utilities\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.086046 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xw99k\" (UniqueName: \"kubernetes.io/projected/e1f282da-749c-42a0-aea4-0ec0587bc5fc-kube-api-access-xw99k\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.086215 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-catalog-content\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.086273 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-utilities\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.102686 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-547dbd544d-hnv7m" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.112296 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xw99k\" (UniqueName: \"kubernetes.io/projected/e1f282da-749c-42a0-aea4-0ec0587bc5fc-kube-api-access-xw99k\") pod \"redhat-marketplace-czz8n\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.186068 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.389347 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz8n"] Dec 11 16:58:35 crc kubenswrapper[5109]: W1211 16:58:35.392512 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1f282da_749c_42a0_aea4_0ec0587bc5fc.slice/crio-a04effc515755b74e9db9059c2283fd4b61969fc02b4b87ddf1ffa1e8a7a7b96 WatchSource:0}: Error finding container a04effc515755b74e9db9059c2283fd4b61969fc02b4b87ddf1ffa1e8a7a7b96: Status 404 returned error can't find the container with id a04effc515755b74e9db9059c2283fd4b61969fc02b4b87ddf1ffa1e8a7a7b96 Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.854891 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-tg6k2"] Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.889043 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tg6k2"] Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.889158 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.891481 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"redhat-operators-dockercfg-9gxlh\"" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.998333 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-catalog-content\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.998493 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk2bt\" (UniqueName: \"kubernetes.io/projected/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-kube-api-access-rk2bt\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:35 crc kubenswrapper[5109]: I1211 16:58:35.998554 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-utilities\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.100216 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rk2bt\" (UniqueName: \"kubernetes.io/projected/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-kube-api-access-rk2bt\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.100365 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-utilities\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.100429 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-catalog-content\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.101092 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-utilities\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.101513 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-catalog-content\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.103347 5109 generic.go:358] "Generic (PLEG): container finished" podID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerID="aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b" exitCode=0 Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.103466 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz8n" event={"ID":"e1f282da-749c-42a0-aea4-0ec0587bc5fc","Type":"ContainerDied","Data":"aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b"} Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.103536 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz8n" event={"ID":"e1f282da-749c-42a0-aea4-0ec0587bc5fc","Type":"ContainerStarted","Data":"a04effc515755b74e9db9059c2283fd4b61969fc02b4b87ddf1ffa1e8a7a7b96"} Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.132965 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk2bt\" (UniqueName: \"kubernetes.io/projected/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-kube-api-access-rk2bt\") pod \"redhat-operators-tg6k2\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.224697 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:36 crc kubenswrapper[5109]: I1211 16:58:36.395067 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-tg6k2"] Dec 11 16:58:36 crc kubenswrapper[5109]: W1211 16:58:36.397941 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ec445aa_ae2d_4903_9a2a_a2e811f43ad6.slice/crio-baf12f7fb5b7588928ac4ed5e3d1f480647f07337981e471c16e01c5271bfbac WatchSource:0}: Error finding container baf12f7fb5b7588928ac4ed5e3d1f480647f07337981e471c16e01c5271bfbac: Status 404 returned error can't find the container with id baf12f7fb5b7588928ac4ed5e3d1f480647f07337981e471c16e01c5271bfbac Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.112690 5109 generic.go:358] "Generic (PLEG): container finished" podID="0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" containerID="de9500dea283e6a2d4a5098b6c5e933e663c2369ed3782562923899943d32580" exitCode=0 Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.112802 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tg6k2" event={"ID":"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6","Type":"ContainerDied","Data":"de9500dea283e6a2d4a5098b6c5e933e663c2369ed3782562923899943d32580"} Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.113442 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tg6k2" event={"ID":"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6","Type":"ContainerStarted","Data":"baf12f7fb5b7588928ac4ed5e3d1f480647f07337981e471c16e01c5271bfbac"} Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.255506 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-pgfnr"] Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.260344 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.263877 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-7cl8d\"" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.271069 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgfnr"] Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.315309 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-utilities\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.315387 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drncs\" (UniqueName: \"kubernetes.io/projected/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-kube-api-access-drncs\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.315434 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-catalog-content\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.417466 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-catalog-content\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.417668 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-utilities\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.417731 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-drncs\" (UniqueName: \"kubernetes.io/projected/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-kube-api-access-drncs\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.418382 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-catalog-content\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.418456 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-utilities\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.453984 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-drncs\" (UniqueName: \"kubernetes.io/projected/9f2ef198-cb50-4dd0-8292-f6c0e68f901f-kube-api-access-drncs\") pod \"certified-operators-pgfnr\" (UID: \"9f2ef198-cb50-4dd0-8292-f6c0e68f901f\") " pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.663393 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:37 crc kubenswrapper[5109]: I1211 16:58:37.934591 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-pgfnr"] Dec 11 16:58:37 crc kubenswrapper[5109]: W1211 16:58:37.945474 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9f2ef198_cb50_4dd0_8292_f6c0e68f901f.slice/crio-3a9ea51c9792f580244c1b7fcff50fdbf6c889ec190802fcdab0a2c48ae8ce74 WatchSource:0}: Error finding container 3a9ea51c9792f580244c1b7fcff50fdbf6c889ec190802fcdab0a2c48ae8ce74: Status 404 returned error can't find the container with id 3a9ea51c9792f580244c1b7fcff50fdbf6c889ec190802fcdab0a2c48ae8ce74 Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.126318 5109 generic.go:358] "Generic (PLEG): container finished" podID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerID="2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319" exitCode=0 Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.126487 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz8n" event={"ID":"e1f282da-749c-42a0-aea4-0ec0587bc5fc","Type":"ContainerDied","Data":"2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319"} Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.132969 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgfnr" event={"ID":"9f2ef198-cb50-4dd0-8292-f6c0e68f901f","Type":"ContainerStarted","Data":"3a9ea51c9792f580244c1b7fcff50fdbf6c889ec190802fcdab0a2c48ae8ce74"} Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.449748 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-q8hjf"] Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.461433 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.464124 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"community-operators-dockercfg-vrd5f\"" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.470472 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q8hjf"] Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.531540 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzgn8\" (UniqueName: \"kubernetes.io/projected/5dbe8975-f527-4e83-acde-9679cdcde3a9-kube-api-access-jzgn8\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.531614 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbe8975-f527-4e83-acde-9679cdcde3a9-utilities\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.531669 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbe8975-f527-4e83-acde-9679cdcde3a9-catalog-content\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.633197 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jzgn8\" (UniqueName: \"kubernetes.io/projected/5dbe8975-f527-4e83-acde-9679cdcde3a9-kube-api-access-jzgn8\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.633251 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbe8975-f527-4e83-acde-9679cdcde3a9-utilities\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.633285 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbe8975-f527-4e83-acde-9679cdcde3a9-catalog-content\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.633882 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5dbe8975-f527-4e83-acde-9679cdcde3a9-catalog-content\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.634252 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5dbe8975-f527-4e83-acde-9679cdcde3a9-utilities\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.652976 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzgn8\" (UniqueName: \"kubernetes.io/projected/5dbe8975-f527-4e83-acde-9679cdcde3a9-kube-api-access-jzgn8\") pod \"community-operators-q8hjf\" (UID: \"5dbe8975-f527-4e83-acde-9679cdcde3a9\") " pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:38 crc kubenswrapper[5109]: I1211 16:58:38.821970 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.042733 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-q8hjf"] Dec 11 16:58:39 crc kubenswrapper[5109]: W1211 16:58:39.048648 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5dbe8975_f527_4e83_acde_9679cdcde3a9.slice/crio-66613336076c583a1c8c6e14c094cef36a38c8b5fd1fab3c253115935b41a0e0 WatchSource:0}: Error finding container 66613336076c583a1c8c6e14c094cef36a38c8b5fd1fab3c253115935b41a0e0: Status 404 returned error can't find the container with id 66613336076c583a1c8c6e14c094cef36a38c8b5fd1fab3c253115935b41a0e0 Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.140364 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz8n" event={"ID":"e1f282da-749c-42a0-aea4-0ec0587bc5fc","Type":"ContainerStarted","Data":"65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c"} Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.142354 5109 generic.go:358] "Generic (PLEG): container finished" podID="9f2ef198-cb50-4dd0-8292-f6c0e68f901f" containerID="384d818e109573c3a411e38fa7964481fb9a2143eac52b6988da4ec16391b1c8" exitCode=0 Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.142439 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgfnr" event={"ID":"9f2ef198-cb50-4dd0-8292-f6c0e68f901f","Type":"ContainerDied","Data":"384d818e109573c3a411e38fa7964481fb9a2143eac52b6988da4ec16391b1c8"} Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.144627 5109 generic.go:358] "Generic (PLEG): container finished" podID="0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" containerID="b027c5c536838006934e7164a78b70425652a1dad8e5d5129ec8f0627af5b463" exitCode=0 Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.144713 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tg6k2" event={"ID":"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6","Type":"ContainerDied","Data":"b027c5c536838006934e7164a78b70425652a1dad8e5d5129ec8f0627af5b463"} Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.146601 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8hjf" event={"ID":"5dbe8975-f527-4e83-acde-9679cdcde3a9","Type":"ContainerStarted","Data":"66613336076c583a1c8c6e14c094cef36a38c8b5fd1fab3c253115935b41a0e0"} Dec 11 16:58:39 crc kubenswrapper[5109]: I1211 16:58:39.163238 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-czz8n" podStartSLOduration=4.210078462 podStartE2EDuration="5.16321962s" podCreationTimestamp="2025-12-11 16:58:34 +0000 UTC" firstStartedPulling="2025-12-11 16:58:36.109259287 +0000 UTC m=+353.788950803" lastFinishedPulling="2025-12-11 16:58:37.062400495 +0000 UTC m=+354.742091961" observedRunningTime="2025-12-11 16:58:39.159841675 +0000 UTC m=+356.839533141" watchObservedRunningTime="2025-12-11 16:58:39.16321962 +0000 UTC m=+356.842911076" Dec 11 16:58:40 crc kubenswrapper[5109]: I1211 16:58:40.156468 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tg6k2" event={"ID":"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6","Type":"ContainerStarted","Data":"0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9"} Dec 11 16:58:40 crc kubenswrapper[5109]: I1211 16:58:40.160807 5109 generic.go:358] "Generic (PLEG): container finished" podID="5dbe8975-f527-4e83-acde-9679cdcde3a9" containerID="7cc94416fbd1bd2aacf37a49d84f1e8c6709715b93bde2b8ef9f6e817caa1666" exitCode=0 Dec 11 16:58:40 crc kubenswrapper[5109]: I1211 16:58:40.162060 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8hjf" event={"ID":"5dbe8975-f527-4e83-acde-9679cdcde3a9","Type":"ContainerDied","Data":"7cc94416fbd1bd2aacf37a49d84f1e8c6709715b93bde2b8ef9f6e817caa1666"} Dec 11 16:58:40 crc kubenswrapper[5109]: I1211 16:58:40.184604 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-tg6k2" podStartSLOduration=4.352461914 podStartE2EDuration="5.184584069s" podCreationTimestamp="2025-12-11 16:58:35 +0000 UTC" firstStartedPulling="2025-12-11 16:58:37.113632148 +0000 UTC m=+354.793323614" lastFinishedPulling="2025-12-11 16:58:37.945754303 +0000 UTC m=+355.625445769" observedRunningTime="2025-12-11 16:58:40.17947098 +0000 UTC m=+357.859162466" watchObservedRunningTime="2025-12-11 16:58:40.184584069 +0000 UTC m=+357.864275545" Dec 11 16:58:41 crc kubenswrapper[5109]: I1211 16:58:41.169237 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8hjf" event={"ID":"5dbe8975-f527-4e83-acde-9679cdcde3a9","Type":"ContainerStarted","Data":"1bc4ac98835a6b45ed4d6b999b9dea8c5a3984b298d361a5465e7abc6d8643c0"} Dec 11 16:58:41 crc kubenswrapper[5109]: I1211 16:58:41.170941 5109 generic.go:358] "Generic (PLEG): container finished" podID="9f2ef198-cb50-4dd0-8292-f6c0e68f901f" containerID="0a708593c587824e935e5d047d091da4b514ef948ca8fe96796f4bfc6db1078d" exitCode=0 Dec 11 16:58:41 crc kubenswrapper[5109]: I1211 16:58:41.171028 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgfnr" event={"ID":"9f2ef198-cb50-4dd0-8292-f6c0e68f901f","Type":"ContainerDied","Data":"0a708593c587824e935e5d047d091da4b514ef948ca8fe96796f4bfc6db1078d"} Dec 11 16:58:42 crc kubenswrapper[5109]: I1211 16:58:42.180964 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-pgfnr" event={"ID":"9f2ef198-cb50-4dd0-8292-f6c0e68f901f","Type":"ContainerStarted","Data":"30537b48f799195e515eeb977cb37ca70d1403531326ba1199ef551505653c80"} Dec 11 16:58:42 crc kubenswrapper[5109]: I1211 16:58:42.183769 5109 generic.go:358] "Generic (PLEG): container finished" podID="5dbe8975-f527-4e83-acde-9679cdcde3a9" containerID="1bc4ac98835a6b45ed4d6b999b9dea8c5a3984b298d361a5465e7abc6d8643c0" exitCode=0 Dec 11 16:58:42 crc kubenswrapper[5109]: I1211 16:58:42.183862 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8hjf" event={"ID":"5dbe8975-f527-4e83-acde-9679cdcde3a9","Type":"ContainerDied","Data":"1bc4ac98835a6b45ed4d6b999b9dea8c5a3984b298d361a5465e7abc6d8643c0"} Dec 11 16:58:42 crc kubenswrapper[5109]: I1211 16:58:42.209458 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-pgfnr" podStartSLOduration=4.303037197 podStartE2EDuration="5.209434107s" podCreationTimestamp="2025-12-11 16:58:37 +0000 UTC" firstStartedPulling="2025-12-11 16:58:39.143063991 +0000 UTC m=+356.822755447" lastFinishedPulling="2025-12-11 16:58:40.049460861 +0000 UTC m=+357.729152357" observedRunningTime="2025-12-11 16:58:42.208234077 +0000 UTC m=+359.887925553" watchObservedRunningTime="2025-12-11 16:58:42.209434107 +0000 UTC m=+359.889125583" Dec 11 16:58:44 crc kubenswrapper[5109]: I1211 16:58:44.201299 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-q8hjf" event={"ID":"5dbe8975-f527-4e83-acde-9679cdcde3a9","Type":"ContainerStarted","Data":"8bd8287a0ab0b0347fe905affbbd6a8d1dc2d1dab19601b43d47c2d38649e218"} Dec 11 16:58:44 crc kubenswrapper[5109]: I1211 16:58:44.229223 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-q8hjf" podStartSLOduration=5.427200672 podStartE2EDuration="6.229200607s" podCreationTimestamp="2025-12-11 16:58:38 +0000 UTC" firstStartedPulling="2025-12-11 16:58:40.162255726 +0000 UTC m=+357.841947202" lastFinishedPulling="2025-12-11 16:58:40.964255651 +0000 UTC m=+358.643947137" observedRunningTime="2025-12-11 16:58:44.225966145 +0000 UTC m=+361.905657631" watchObservedRunningTime="2025-12-11 16:58:44.229200607 +0000 UTC m=+361.908892113" Dec 11 16:58:45 crc kubenswrapper[5109]: I1211 16:58:45.186878 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:45 crc kubenswrapper[5109]: I1211 16:58:45.187138 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:45 crc kubenswrapper[5109]: I1211 16:58:45.231509 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:45 crc kubenswrapper[5109]: I1211 16:58:45.262820 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 16:58:46 crc kubenswrapper[5109]: I1211 16:58:46.225665 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:46 crc kubenswrapper[5109]: I1211 16:58:46.225841 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:46 crc kubenswrapper[5109]: I1211 16:58:46.292113 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:47 crc kubenswrapper[5109]: I1211 16:58:47.274436 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 16:58:47 crc kubenswrapper[5109]: I1211 16:58:47.664410 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:47 crc kubenswrapper[5109]: I1211 16:58:47.664462 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:47 crc kubenswrapper[5109]: I1211 16:58:47.719457 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:48 crc kubenswrapper[5109]: I1211 16:58:48.269016 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-pgfnr" Dec 11 16:58:48 crc kubenswrapper[5109]: I1211 16:58:48.822375 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:48 crc kubenswrapper[5109]: I1211 16:58:48.823474 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:48 crc kubenswrapper[5109]: I1211 16:58:48.861714 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-q8hjf" Dec 11 16:58:49 crc kubenswrapper[5109]: I1211 16:58:49.268581 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-q8hjf" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.204612 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5"] Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.216030 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5"] Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.216194 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.219329 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.222049 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.338171 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/939540de-b185-4b5a-a5ed-3f257a0c9010-secret-volume\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.338293 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zsv9\" (UniqueName: \"kubernetes.io/projected/939540de-b185-4b5a-a5ed-3f257a0c9010-kube-api-access-5zsv9\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.338341 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/939540de-b185-4b5a-a5ed-3f257a0c9010-config-volume\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.440199 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/939540de-b185-4b5a-a5ed-3f257a0c9010-secret-volume\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.440269 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-5zsv9\" (UniqueName: \"kubernetes.io/projected/939540de-b185-4b5a-a5ed-3f257a0c9010-kube-api-access-5zsv9\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.440300 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/939540de-b185-4b5a-a5ed-3f257a0c9010-config-volume\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.442164 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/939540de-b185-4b5a-a5ed-3f257a0c9010-config-volume\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.451585 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/939540de-b185-4b5a-a5ed-3f257a0c9010-secret-volume\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.472778 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zsv9\" (UniqueName: \"kubernetes.io/projected/939540de-b185-4b5a-a5ed-3f257a0c9010-kube-api-access-5zsv9\") pod \"collect-profiles-29424540-zcdq5\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.539182 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:00 crc kubenswrapper[5109]: I1211 17:00:00.841849 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5"] Dec 11 17:00:00 crc kubenswrapper[5109]: W1211 17:00:00.848538 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod939540de_b185_4b5a_a5ed_3f257a0c9010.slice/crio-1b255495a91166847963ac5f6091584f0419677e16c3aea8c72d9c747e801e7f WatchSource:0}: Error finding container 1b255495a91166847963ac5f6091584f0419677e16c3aea8c72d9c747e801e7f: Status 404 returned error can't find the container with id 1b255495a91166847963ac5f6091584f0419677e16c3aea8c72d9c747e801e7f Dec 11 17:00:01 crc kubenswrapper[5109]: I1211 17:00:01.765001 5109 generic.go:358] "Generic (PLEG): container finished" podID="939540de-b185-4b5a-a5ed-3f257a0c9010" containerID="abc4699707fd9775cbad74b88c3679c3b0efa02c441ec72b27987c35109dfa6e" exitCode=0 Dec 11 17:00:01 crc kubenswrapper[5109]: I1211 17:00:01.765064 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" event={"ID":"939540de-b185-4b5a-a5ed-3f257a0c9010","Type":"ContainerDied","Data":"abc4699707fd9775cbad74b88c3679c3b0efa02c441ec72b27987c35109dfa6e"} Dec 11 17:00:01 crc kubenswrapper[5109]: I1211 17:00:01.765355 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" event={"ID":"939540de-b185-4b5a-a5ed-3f257a0c9010","Type":"ContainerStarted","Data":"1b255495a91166847963ac5f6091584f0419677e16c3aea8c72d9c747e801e7f"} Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.053266 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.178782 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/939540de-b185-4b5a-a5ed-3f257a0c9010-config-volume\") pod \"939540de-b185-4b5a-a5ed-3f257a0c9010\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.178970 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/939540de-b185-4b5a-a5ed-3f257a0c9010-secret-volume\") pod \"939540de-b185-4b5a-a5ed-3f257a0c9010\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.179086 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zsv9\" (UniqueName: \"kubernetes.io/projected/939540de-b185-4b5a-a5ed-3f257a0c9010-kube-api-access-5zsv9\") pod \"939540de-b185-4b5a-a5ed-3f257a0c9010\" (UID: \"939540de-b185-4b5a-a5ed-3f257a0c9010\") " Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.179912 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/939540de-b185-4b5a-a5ed-3f257a0c9010-config-volume" (OuterVolumeSpecName: "config-volume") pod "939540de-b185-4b5a-a5ed-3f257a0c9010" (UID: "939540de-b185-4b5a-a5ed-3f257a0c9010"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.189035 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/939540de-b185-4b5a-a5ed-3f257a0c9010-kube-api-access-5zsv9" (OuterVolumeSpecName: "kube-api-access-5zsv9") pod "939540de-b185-4b5a-a5ed-3f257a0c9010" (UID: "939540de-b185-4b5a-a5ed-3f257a0c9010"). InnerVolumeSpecName "kube-api-access-5zsv9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.189942 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/939540de-b185-4b5a-a5ed-3f257a0c9010-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "939540de-b185-4b5a-a5ed-3f257a0c9010" (UID: "939540de-b185-4b5a-a5ed-3f257a0c9010"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.280895 5109 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/939540de-b185-4b5a-a5ed-3f257a0c9010-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.280943 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-5zsv9\" (UniqueName: \"kubernetes.io/projected/939540de-b185-4b5a-a5ed-3f257a0c9010-kube-api-access-5zsv9\") on node \"crc\" DevicePath \"\"" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.280963 5109 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/939540de-b185-4b5a-a5ed-3f257a0c9010-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.781982 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" event={"ID":"939540de-b185-4b5a-a5ed-3f257a0c9010","Type":"ContainerDied","Data":"1b255495a91166847963ac5f6091584f0419677e16c3aea8c72d9c747e801e7f"} Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.782039 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b255495a91166847963ac5f6091584f0419677e16c3aea8c72d9c747e801e7f" Dec 11 17:00:03 crc kubenswrapper[5109]: I1211 17:00:03.781997 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424540-zcdq5" Dec 11 17:00:07 crc kubenswrapper[5109]: I1211 17:00:07.762043 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:00:07 crc kubenswrapper[5109]: I1211 17:00:07.762391 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:00:37 crc kubenswrapper[5109]: I1211 17:00:37.762951 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:00:37 crc kubenswrapper[5109]: I1211 17:00:37.763784 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:01:07 crc kubenswrapper[5109]: I1211 17:01:07.762183 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:01:07 crc kubenswrapper[5109]: I1211 17:01:07.762865 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:01:07 crc kubenswrapper[5109]: I1211 17:01:07.762926 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 17:01:07 crc kubenswrapper[5109]: I1211 17:01:07.763899 5109 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfe59f5af7913b6af69c36c59fa1989b46d095e49ddef2e9386892be6189ace2"} pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 17:01:07 crc kubenswrapper[5109]: I1211 17:01:07.764032 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" containerID="cri-o://dfe59f5af7913b6af69c36c59fa1989b46d095e49ddef2e9386892be6189ace2" gracePeriod=600 Dec 11 17:01:08 crc kubenswrapper[5109]: I1211 17:01:08.227431 5109 generic.go:358] "Generic (PLEG): container finished" podID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerID="dfe59f5af7913b6af69c36c59fa1989b46d095e49ddef2e9386892be6189ace2" exitCode=0 Dec 11 17:01:08 crc kubenswrapper[5109]: I1211 17:01:08.227547 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerDied","Data":"dfe59f5af7913b6af69c36c59fa1989b46d095e49ddef2e9386892be6189ace2"} Dec 11 17:01:08 crc kubenswrapper[5109]: I1211 17:01:08.227963 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"5113406771ea4df98ffaeb4cf5ec00ae40484fd2ce8f067a051e90a693151965"} Dec 11 17:01:08 crc kubenswrapper[5109]: I1211 17:01:08.228005 5109 scope.go:117] "RemoveContainer" containerID="f099a425a82983d885b200b73fcf89b1dba37bc85eba89c7f84a71efcc7178c7" Dec 11 17:01:43 crc kubenswrapper[5109]: I1211 17:01:43.741535 5109 scope.go:117] "RemoveContainer" containerID="e57f87ee47ac6a19a1505c54775ea70c7f2ef80da763216df55ef895048d5412" Dec 11 17:02:43 crc kubenswrapper[5109]: I1211 17:02:43.139205 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 17:02:43 crc kubenswrapper[5109]: I1211 17:02:43.142823 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 17:02:43 crc kubenswrapper[5109]: I1211 17:02:43.215575 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 17:02:43 crc kubenswrapper[5109]: I1211 17:02:43.217474 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 17:03:02 crc kubenswrapper[5109]: I1211 17:03:02.299453 5109 ???:1] "http: TLS handshake error from 192.168.126.11:33960: no serving certificate available for the kubelet" Dec 11 17:03:37 crc kubenswrapper[5109]: I1211 17:03:37.761617 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:03:37 crc kubenswrapper[5109]: I1211 17:03:37.762237 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:04:07 crc kubenswrapper[5109]: I1211 17:04:07.762087 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:04:07 crc kubenswrapper[5109]: I1211 17:04:07.762684 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:04:37 crc kubenswrapper[5109]: I1211 17:04:37.762183 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:04:37 crc kubenswrapper[5109]: I1211 17:04:37.762733 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:04:37 crc kubenswrapper[5109]: I1211 17:04:37.762830 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 17:04:37 crc kubenswrapper[5109]: I1211 17:04:37.763586 5109 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5113406771ea4df98ffaeb4cf5ec00ae40484fd2ce8f067a051e90a693151965"} pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 17:04:37 crc kubenswrapper[5109]: I1211 17:04:37.763673 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" containerID="cri-o://5113406771ea4df98ffaeb4cf5ec00ae40484fd2ce8f067a051e90a693151965" gracePeriod=600 Dec 11 17:04:37 crc kubenswrapper[5109]: I1211 17:04:37.904667 5109 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 17:04:38 crc kubenswrapper[5109]: I1211 17:04:38.752098 5109 generic.go:358] "Generic (PLEG): container finished" podID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerID="5113406771ea4df98ffaeb4cf5ec00ae40484fd2ce8f067a051e90a693151965" exitCode=0 Dec 11 17:04:38 crc kubenswrapper[5109]: I1211 17:04:38.752186 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerDied","Data":"5113406771ea4df98ffaeb4cf5ec00ae40484fd2ce8f067a051e90a693151965"} Dec 11 17:04:38 crc kubenswrapper[5109]: I1211 17:04:38.752643 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"1f3b25744dcfda371142af697057ed29ab8b623200bea5beeda0e00d433c8883"} Dec 11 17:04:38 crc kubenswrapper[5109]: I1211 17:04:38.752673 5109 scope.go:117] "RemoveContainer" containerID="dfe59f5af7913b6af69c36c59fa1989b46d095e49ddef2e9386892be6189ace2" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.212062 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n"] Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.212805 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="kube-rbac-proxy" containerID="cri-o://82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.213272 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="ovnkube-cluster-manager" containerID="cri-o://ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.416178 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wfxmd"] Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.416775 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-controller" containerID="cri-o://ddb82e106ea206c6481e0e7996db55518f439ec5ea77c08b9b391f2eaced3ecf" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.416852 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="sbdb" containerID="cri-o://a9ecd830813d47e96969244bfd5812403f28e62d92f8d68d84eb4dae08215ad2" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.416918 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="northd" containerID="cri-o://c59ff5a852839e949601b2617bdae48e2624b8bef3e61979cbada802763747f6" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.416950 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-acl-logging" containerID="cri-o://068013b5c151d23b07bff0a7c4f87ba16ae9050d22993e0cb0474e1e32d7fe9b" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.416985 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="nbdb" containerID="cri-o://6f48c12cd8d9d7f9da38db4dd4293bdba9bd084e234c25324a74d281e52162eb" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.417018 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://f56e7322033f70f60b9b525d77378a394971fe110d41e3f8ca772fc8ae86d4ed" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.417049 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-node" containerID="cri-o://99d2321957866d3a56bdda8beca1e337cca93f57ad8fcdcae93e8787590c10a7" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.448826 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovnkube-controller" containerID="cri-o://052d8728ba15ac15e54b397f9633ec810798f5f74e7189ec98767d0afdcf916a" gracePeriod=30 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.462477 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.486138 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f"] Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.486670 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="kube-rbac-proxy" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.486687 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="kube-rbac-proxy" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.486701 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="939540de-b185-4b5a-a5ed-3f257a0c9010" containerName="collect-profiles" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.486708 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="939540de-b185-4b5a-a5ed-3f257a0c9010" containerName="collect-profiles" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.486720 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="ovnkube-cluster-manager" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.486726 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="ovnkube-cluster-manager" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.487546 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="939540de-b185-4b5a-a5ed-3f257a0c9010" containerName="collect-profiles" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.487569 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="ovnkube-cluster-manager" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.487585 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerName="kube-rbac-proxy" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.493034 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.513005 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-env-overrides\") pod \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.513067 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovnkube-config\") pod \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.513100 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovn-control-plane-metrics-cert\") pod \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.513192 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4fpt\" (UniqueName: \"kubernetes.io/projected/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-kube-api-access-c4fpt\") pod \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\" (UID: \"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9\") " Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.513629 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" (UID: "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.513863 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" (UID: "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.521010 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" (UID: "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.521303 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-kube-api-access-c4fpt" (OuterVolumeSpecName: "kube-api-access-c4fpt") pod "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" (UID: "5f6bde63-48d8-4726-b7d6-0f36f7cf47f9"). InnerVolumeSpecName "kube-api-access-c4fpt". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614309 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614352 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614415 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l4gxs\" (UniqueName: \"kubernetes.io/projected/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-kube-api-access-l4gxs\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614462 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614535 5109 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614551 5109 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614563 5109 reconciler_common.go:299] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.614578 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-c4fpt\" (UniqueName: \"kubernetes.io/projected/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9-kube-api-access-c4fpt\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.716199 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.716375 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.717211 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.717554 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-l4gxs\" (UniqueName: \"kubernetes.io/projected/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-kube-api-access-l4gxs\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.718054 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-env-overrides\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.718432 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-ovnkube-config\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.722776 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.736174 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-l4gxs\" (UniqueName: \"kubernetes.io/projected/4c4c9d34-3d7e-4368-a7b6-e867a3b54542-kube-api-access-l4gxs\") pod \"ovnkube-control-plane-97c9b6c48-wjk5f\" (UID: \"4c4c9d34-3d7e-4368-a7b6-e867a3b54542\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.787114 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hx2kb_7c45b4ae-02aa-4614-977b-544a9e303bdf/kube-multus/0.log" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.787368 5109 generic.go:358] "Generic (PLEG): container finished" podID="7c45b4ae-02aa-4614-977b-544a9e303bdf" containerID="745ca6fee05059f7ae1cc1c5da3bbab828e24840a54cf61bb26296bf93ce4733" exitCode=2 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.787467 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hx2kb" event={"ID":"7c45b4ae-02aa-4614-977b-544a9e303bdf","Type":"ContainerDied","Data":"745ca6fee05059f7ae1cc1c5da3bbab828e24840a54cf61bb26296bf93ce4733"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.788556 5109 scope.go:117] "RemoveContainer" containerID="745ca6fee05059f7ae1cc1c5da3bbab828e24840a54cf61bb26296bf93ce4733" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.795148 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wfxmd_561da642-b358-4ddc-86ee-669164d7b3c5/ovn-acl-logging/0.log" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796021 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wfxmd_561da642-b358-4ddc-86ee-669164d7b3c5/ovn-controller/0.log" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796695 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="052d8728ba15ac15e54b397f9633ec810798f5f74e7189ec98767d0afdcf916a" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796772 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="a9ecd830813d47e96969244bfd5812403f28e62d92f8d68d84eb4dae08215ad2" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796794 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="6f48c12cd8d9d7f9da38db4dd4293bdba9bd084e234c25324a74d281e52162eb" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796810 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="c59ff5a852839e949601b2617bdae48e2624b8bef3e61979cbada802763747f6" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796825 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="f56e7322033f70f60b9b525d77378a394971fe110d41e3f8ca772fc8ae86d4ed" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796842 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="99d2321957866d3a56bdda8beca1e337cca93f57ad8fcdcae93e8787590c10a7" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796859 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="068013b5c151d23b07bff0a7c4f87ba16ae9050d22993e0cb0474e1e32d7fe9b" exitCode=143 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.796875 5109 generic.go:358] "Generic (PLEG): container finished" podID="561da642-b358-4ddc-86ee-669164d7b3c5" containerID="ddb82e106ea206c6481e0e7996db55518f439ec5ea77c08b9b391f2eaced3ecf" exitCode=143 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797054 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"052d8728ba15ac15e54b397f9633ec810798f5f74e7189ec98767d0afdcf916a"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797105 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"a9ecd830813d47e96969244bfd5812403f28e62d92f8d68d84eb4dae08215ad2"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797129 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"6f48c12cd8d9d7f9da38db4dd4293bdba9bd084e234c25324a74d281e52162eb"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797149 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"c59ff5a852839e949601b2617bdae48e2624b8bef3e61979cbada802763747f6"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797166 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"f56e7322033f70f60b9b525d77378a394971fe110d41e3f8ca772fc8ae86d4ed"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797182 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"99d2321957866d3a56bdda8beca1e337cca93f57ad8fcdcae93e8787590c10a7"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797201 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"068013b5c151d23b07bff0a7c4f87ba16ae9050d22993e0cb0474e1e32d7fe9b"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.797218 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"ddb82e106ea206c6481e0e7996db55518f439ec5ea77c08b9b391f2eaced3ecf"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.804873 5109 generic.go:358] "Generic (PLEG): container finished" podID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerID="ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.804917 5109 generic.go:358] "Generic (PLEG): container finished" podID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" containerID="82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498" exitCode=0 Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.804977 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" event={"ID":"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9","Type":"ContainerDied","Data":"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.805084 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" event={"ID":"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9","Type":"ContainerDied","Data":"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.805108 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" event={"ID":"5f6bde63-48d8-4726-b7d6-0f36f7cf47f9","Type":"ContainerDied","Data":"f40c60bdf53415aae65189c284cbebca1497a09cbfeb3296bd29d5af60ad6ecb"} Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.805138 5109 scope.go:117] "RemoveContainer" containerID="ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.805015 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.834636 5109 scope.go:117] "RemoveContainer" containerID="82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.847578 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.854005 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n"] Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.861786 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-57b78d8988-dzv7n"] Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.877175 5109 scope.go:117] "RemoveContainer" containerID="ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0" Dec 11 17:04:41 crc kubenswrapper[5109]: E1211 17:04:41.877713 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0\": container with ID starting with ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0 not found: ID does not exist" containerID="ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.877798 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0"} err="failed to get container status \"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0\": rpc error: code = NotFound desc = could not find container \"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0\": container with ID starting with ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0 not found: ID does not exist" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.877837 5109 scope.go:117] "RemoveContainer" containerID="82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498" Dec 11 17:04:41 crc kubenswrapper[5109]: E1211 17:04:41.879491 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498\": container with ID starting with 82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498 not found: ID does not exist" containerID="82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.879526 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498"} err="failed to get container status \"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498\": rpc error: code = NotFound desc = could not find container \"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498\": container with ID starting with 82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498 not found: ID does not exist" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.879551 5109 scope.go:117] "RemoveContainer" containerID="ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.880001 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0"} err="failed to get container status \"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0\": rpc error: code = NotFound desc = could not find container \"ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0\": container with ID starting with ebcdeecba38f0e015adc2d94019559cfae2ca6bab7ec9e40b3d256d97cb522f0 not found: ID does not exist" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.880026 5109 scope.go:117] "RemoveContainer" containerID="82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498" Dec 11 17:04:41 crc kubenswrapper[5109]: I1211 17:04:41.880464 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498"} err="failed to get container status \"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498\": rpc error: code = NotFound desc = could not find container \"82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498\": container with ID starting with 82d8fad494147a0e97628339de8066fcef78933d8de216e575a52eb71e03c498 not found: ID does not exist" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.104797 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wfxmd_561da642-b358-4ddc-86ee-669164d7b3c5/ovn-acl-logging/0.log" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.105517 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wfxmd_561da642-b358-4ddc-86ee-669164d7b3c5/ovn-controller/0.log" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.106160 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197042 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-9f5js"] Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197763 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovnkube-controller" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197783 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovnkube-controller" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197803 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="sbdb" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197810 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="sbdb" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197823 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-node" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197831 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-node" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197844 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="northd" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197851 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="northd" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197863 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197872 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197885 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-controller" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197892 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-controller" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197908 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="nbdb" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197915 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="nbdb" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197929 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kubecfg-setup" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197938 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kubecfg-setup" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197955 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-acl-logging" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.197963 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-acl-logging" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198078 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-ovn-metrics" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198094 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="nbdb" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198105 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovnkube-controller" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198116 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="sbdb" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198124 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-controller" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198135 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="ovn-acl-logging" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198146 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="kube-rbac-proxy-node" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.198156 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" containerName="northd" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.212192 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224688 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/561da642-b358-4ddc-86ee-669164d7b3c5-ovn-node-metrics-cert\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224753 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-ovn\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224793 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-systemd-units\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224825 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-config\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224855 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46pxq\" (UniqueName: \"kubernetes.io/projected/561da642-b358-4ddc-86ee-669164d7b3c5-kube-api-access-46pxq\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224877 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-netns\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224904 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-ovn-kubernetes\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224925 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-openvswitch\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224948 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-etc-openvswitch\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.224987 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-kubelet\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225041 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-script-lib\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225240 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-env-overrides\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225257 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-slash\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225274 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-netd\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225332 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225375 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-systemd\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225405 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-bin\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225433 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-node-log\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225458 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-log-socket\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225463 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225492 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-var-lib-openvswitch\") pod \"561da642-b358-4ddc-86ee-669164d7b3c5\" (UID: \"561da642-b358-4ddc-86ee-669164d7b3c5\") " Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225537 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225572 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225594 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225685 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225764 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-slash" (OuterVolumeSpecName: "host-slash") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.225820 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226314 5109 reconciler_common.go:299] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226331 5109 reconciler_common.go:299] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226342 5109 reconciler_common.go:299] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226351 5109 reconciler_common.go:299] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226361 5109 reconciler_common.go:299] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226369 5109 reconciler_common.go:299] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-slash\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226377 5109 reconciler_common.go:299] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226878 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226889 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226923 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226922 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226944 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226969 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-node-log" (OuterVolumeSpecName: "node-log") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226990 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.226986 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.227032 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.227143 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-log-socket" (OuterVolumeSpecName: "log-socket") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.233636 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561da642-b358-4ddc-86ee-669164d7b3c5-kube-api-access-46pxq" (OuterVolumeSpecName: "kube-api-access-46pxq") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "kube-api-access-46pxq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.234701 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561da642-b358-4ddc-86ee-669164d7b3c5-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.244137 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "561da642-b358-4ddc-86ee-669164d7b3c5" (UID: "561da642-b358-4ddc-86ee-669164d7b3c5"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGIDValue "" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329021 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-run-netns\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329082 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-cni-netd\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329154 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-run-ovn-kubernetes\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329197 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovnkube-config\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329252 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329289 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-systemd\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329322 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-etc-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329353 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-ovn\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329386 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-log-socket\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329444 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-node-log\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329487 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-kubelet\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329518 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-env-overrides\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329561 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329643 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxsmt\" (UniqueName: \"kubernetes.io/projected/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-kube-api-access-hxsmt\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329686 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovnkube-script-lib\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329723 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-var-lib-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329931 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-cni-bin\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.329980 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-systemd-units\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330008 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovn-node-metrics-cert\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330031 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-slash\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330093 5109 reconciler_common.go:299] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330107 5109 reconciler_common.go:299] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330119 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-46pxq\" (UniqueName: \"kubernetes.io/projected/561da642-b358-4ddc-86ee-669164d7b3c5-kube-api-access-46pxq\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330132 5109 reconciler_common.go:299] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330143 5109 reconciler_common.go:299] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330155 5109 reconciler_common.go:299] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330167 5109 reconciler_common.go:299] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/561da642-b358-4ddc-86ee-669164d7b3c5-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330181 5109 reconciler_common.go:299] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330192 5109 reconciler_common.go:299] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330204 5109 reconciler_common.go:299] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-node-log\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330214 5109 reconciler_common.go:299] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-log-socket\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330227 5109 reconciler_common.go:299] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/561da642-b358-4ddc-86ee-669164d7b3c5-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.330239 5109 reconciler_common.go:299] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/561da642-b358-4ddc-86ee-669164d7b3c5-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.430863 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431165 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hxsmt\" (UniqueName: \"kubernetes.io/projected/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-kube-api-access-hxsmt\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431010 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431190 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovnkube-script-lib\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431362 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-var-lib-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431456 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-cni-bin\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431515 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-systemd-units\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431562 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovn-node-metrics-cert\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431567 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-var-lib-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431669 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-cni-bin\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431628 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-systemd-units\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431679 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-slash\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431613 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-slash\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431843 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-run-netns\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431867 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-cni-netd\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431977 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-run-netns\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.431979 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-run-ovn-kubernetes\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432025 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-run-ovn-kubernetes\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432038 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovnkube-script-lib\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432056 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-cni-netd\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432067 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovnkube-config\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432149 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432194 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-systemd\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432200 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432221 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-etc-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432241 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-ovn\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432260 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-systemd\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432267 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-log-socket\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432296 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-run-ovn\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432331 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-node-log\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432372 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-log-socket\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432387 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-kubelet\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432336 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-etc-openvswitch\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432415 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-node-log\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432449 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-host-kubelet\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.432453 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-env-overrides\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.433382 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovnkube-config\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.433399 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-env-overrides\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.437539 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-ovn-node-metrics-cert\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.458048 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxsmt\" (UniqueName: \"kubernetes.io/projected/64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3-kube-api-access-hxsmt\") pod \"ovnkube-node-9f5js\" (UID: \"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3\") " pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.534789 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:42 crc kubenswrapper[5109]: W1211 17:04:42.570176 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64de8f9a_f5b7_49ec_ad67_a6e3687f6ca3.slice/crio-f463a420e6c5f449a5ab7d004481fd8d46840584c47f430936c60ab486994bd2 WatchSource:0}: Error finding container f463a420e6c5f449a5ab7d004481fd8d46840584c47f430936c60ab486994bd2: Status 404 returned error can't find the container with id f463a420e6c5f449a5ab7d004481fd8d46840584c47f430936c60ab486994bd2 Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.818506 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wfxmd_561da642-b358-4ddc-86ee-669164d7b3c5/ovn-acl-logging/0.log" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.819532 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-wfxmd_561da642-b358-4ddc-86ee-669164d7b3c5/ovn-controller/0.log" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.820267 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" event={"ID":"561da642-b358-4ddc-86ee-669164d7b3c5","Type":"ContainerDied","Data":"c36f1d88fedba6cd1a52d696a396a0a6177473d575d2e49ea1426299b95ac0f2"} Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.820423 5109 scope.go:117] "RemoveContainer" containerID="052d8728ba15ac15e54b397f9633ec810798f5f74e7189ec98767d0afdcf916a" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.820867 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-wfxmd" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.825706 5109 generic.go:358] "Generic (PLEG): container finished" podID="64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3" containerID="048a7ed003dd59bd66b767d81cbedaf7dd80f257e5d1903a2e18766e785774ec" exitCode=0 Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.825816 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerDied","Data":"048a7ed003dd59bd66b767d81cbedaf7dd80f257e5d1903a2e18766e785774ec"} Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.825845 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"f463a420e6c5f449a5ab7d004481fd8d46840584c47f430936c60ab486994bd2"} Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.841474 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" event={"ID":"4c4c9d34-3d7e-4368-a7b6-e867a3b54542","Type":"ContainerStarted","Data":"37e7e5fd6a1930153e82b5b1cec783a927ce7acfebb45c6f8baccf333d46b124"} Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.841513 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" event={"ID":"4c4c9d34-3d7e-4368-a7b6-e867a3b54542","Type":"ContainerStarted","Data":"8947bc8fa38d2e9406d508c9bbdf4a33fbd0b45810d47f68d2e564ed667ab525"} Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.841528 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" event={"ID":"4c4c9d34-3d7e-4368-a7b6-e867a3b54542","Type":"ContainerStarted","Data":"d725830f6cc2a8ffc8afd1701c2c524432b8a00f3ce9c4e727a12c90ce880a14"} Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.846185 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hx2kb_7c45b4ae-02aa-4614-977b-544a9e303bdf/kube-multus/0.log" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.846305 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-hx2kb" event={"ID":"7c45b4ae-02aa-4614-977b-544a9e303bdf","Type":"ContainerStarted","Data":"adffe609bf461ec10580ac2916cd6f29451bd914df1579884c46aef11a2a8bc8"} Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.857310 5109 scope.go:117] "RemoveContainer" containerID="a9ecd830813d47e96969244bfd5812403f28e62d92f8d68d84eb4dae08215ad2" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.891106 5109 scope.go:117] "RemoveContainer" containerID="6f48c12cd8d9d7f9da38db4dd4293bdba9bd084e234c25324a74d281e52162eb" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.909427 5109 scope.go:117] "RemoveContainer" containerID="c59ff5a852839e949601b2617bdae48e2624b8bef3e61979cbada802763747f6" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.916939 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-97c9b6c48-wjk5f" podStartSLOduration=1.916921745 podStartE2EDuration="1.916921745s" podCreationTimestamp="2025-12-11 17:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 17:04:42.890218204 +0000 UTC m=+720.569909710" watchObservedRunningTime="2025-12-11 17:04:42.916921745 +0000 UTC m=+720.596613211" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.918069 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f6bde63-48d8-4726-b7d6-0f36f7cf47f9" path="/var/lib/kubelet/pods/5f6bde63-48d8-4726-b7d6-0f36f7cf47f9/volumes" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.937037 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wfxmd"] Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.937120 5109 scope.go:117] "RemoveContainer" containerID="f56e7322033f70f60b9b525d77378a394971fe110d41e3f8ca772fc8ae86d4ed" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.942859 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-wfxmd"] Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.962524 5109 scope.go:117] "RemoveContainer" containerID="99d2321957866d3a56bdda8beca1e337cca93f57ad8fcdcae93e8787590c10a7" Dec 11 17:04:42 crc kubenswrapper[5109]: I1211 17:04:42.981324 5109 scope.go:117] "RemoveContainer" containerID="068013b5c151d23b07bff0a7c4f87ba16ae9050d22993e0cb0474e1e32d7fe9b" Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.005125 5109 scope.go:117] "RemoveContainer" containerID="ddb82e106ea206c6481e0e7996db55518f439ec5ea77c08b9b391f2eaced3ecf" Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.022986 5109 scope.go:117] "RemoveContainer" containerID="73bb9591f8cc88cb2850f8019795fc95ce1b6c4d1c220ccb45d0b565dafeb35a" Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.878564 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"f605ed431a8c93ec6fda97545c200b8f46c2682d9d4e2a6ba58b4d49621911aa"} Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.878866 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"0abccdc698635e5794e3be183d344c1013a8f4325f494080505e204b50396df4"} Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.878880 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"5eaf4b738baa2f7a63dc9e9bf7278ac4054798e55b02fe135b0a8de0404f8a2b"} Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.878893 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"f002df057b3f756976844edaf23a50c214f74b12bfaa2d16525ad853fccf4e31"} Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.878904 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"af18375b9b9257d78c1e0220942d6ba4e428d4388df37e2ba5b9ff19705dba5b"} Dec 11 17:04:43 crc kubenswrapper[5109]: I1211 17:04:43.878916 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"ac3ed5af9943331a4515d0248096c3ce9931b3dd4e33dec0507ea2271c3f4573"} Dec 11 17:04:44 crc kubenswrapper[5109]: I1211 17:04:44.911142 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561da642-b358-4ddc-86ee-669164d7b3c5" path="/var/lib/kubelet/pods/561da642-b358-4ddc-86ee-669164d7b3c5/volumes" Dec 11 17:04:46 crc kubenswrapper[5109]: I1211 17:04:46.909897 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"83bf252be110390a80cf5cac37ba97226784f4bc7e9d98e5b553e5fd6ef8d4de"} Dec 11 17:04:50 crc kubenswrapper[5109]: I1211 17:04:50.934290 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" event={"ID":"64de8f9a-f5b7-49ec-ad67-a6e3687f6ca3","Type":"ContainerStarted","Data":"63e6e5492d6d81174e9d22010f298d7c4f7432cfc670b43699be9ce62cfe5ddd"} Dec 11 17:04:50 crc kubenswrapper[5109]: I1211 17:04:50.934949 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:50 crc kubenswrapper[5109]: I1211 17:04:50.979105 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" podStartSLOduration=8.979076311 podStartE2EDuration="8.979076311s" podCreationTimestamp="2025-12-11 17:04:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 17:04:50.972609743 +0000 UTC m=+728.652301269" watchObservedRunningTime="2025-12-11 17:04:50.979076311 +0000 UTC m=+728.658767827" Dec 11 17:04:50 crc kubenswrapper[5109]: I1211 17:04:50.991108 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:51 crc kubenswrapper[5109]: I1211 17:04:51.939163 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:51 crc kubenswrapper[5109]: I1211 17:04:51.939675 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:04:51 crc kubenswrapper[5109]: I1211 17:04:51.975248 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.552449 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-996cl"] Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.568510 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.568806 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-996cl"] Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.735428 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trbd5\" (UniqueName: \"kubernetes.io/projected/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-kube-api-access-trbd5\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.735576 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-utilities\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.735763 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-catalog-content\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.836901 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-utilities\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.837004 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-catalog-content\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.837055 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-trbd5\" (UniqueName: \"kubernetes.io/projected/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-kube-api-access-trbd5\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.837717 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-utilities\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.837840 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-catalog-content\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.876284 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-trbd5\" (UniqueName: \"kubernetes.io/projected/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-kube-api-access-trbd5\") pod \"redhat-operators-996cl\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:22 crc kubenswrapper[5109]: I1211 17:05:22.899402 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:23 crc kubenswrapper[5109]: I1211 17:05:23.375776 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-996cl"] Dec 11 17:05:23 crc kubenswrapper[5109]: I1211 17:05:23.985994 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-9f5js" Dec 11 17:05:24 crc kubenswrapper[5109]: I1211 17:05:24.195152 5109 generic.go:358] "Generic (PLEG): container finished" podID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerID="bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5" exitCode=0 Dec 11 17:05:24 crc kubenswrapper[5109]: I1211 17:05:24.195280 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-996cl" event={"ID":"ad7a318c-68d9-48a9-bc3d-c7e99b64b186","Type":"ContainerDied","Data":"bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5"} Dec 11 17:05:24 crc kubenswrapper[5109]: I1211 17:05:24.195567 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-996cl" event={"ID":"ad7a318c-68d9-48a9-bc3d-c7e99b64b186","Type":"ContainerStarted","Data":"0e0617fe183cd2d058526738bd286209567449f17314daa28c62f7dc858aa364"} Dec 11 17:05:26 crc kubenswrapper[5109]: I1211 17:05:26.209348 5109 generic.go:358] "Generic (PLEG): container finished" podID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerID="23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695" exitCode=0 Dec 11 17:05:26 crc kubenswrapper[5109]: I1211 17:05:26.209418 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-996cl" event={"ID":"ad7a318c-68d9-48a9-bc3d-c7e99b64b186","Type":"ContainerDied","Data":"23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695"} Dec 11 17:05:27 crc kubenswrapper[5109]: I1211 17:05:27.220995 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-996cl" event={"ID":"ad7a318c-68d9-48a9-bc3d-c7e99b64b186","Type":"ContainerStarted","Data":"af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538"} Dec 11 17:05:27 crc kubenswrapper[5109]: I1211 17:05:27.256983 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-996cl" podStartSLOduration=3.887916553 podStartE2EDuration="5.256951504s" podCreationTimestamp="2025-12-11 17:05:22 +0000 UTC" firstStartedPulling="2025-12-11 17:05:24.196300499 +0000 UTC m=+761.875991965" lastFinishedPulling="2025-12-11 17:05:25.56533544 +0000 UTC m=+763.245026916" observedRunningTime="2025-12-11 17:05:27.250939397 +0000 UTC m=+764.930630953" watchObservedRunningTime="2025-12-11 17:05:27.256951504 +0000 UTC m=+764.936643010" Dec 11 17:05:32 crc kubenswrapper[5109]: I1211 17:05:32.908511 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:32 crc kubenswrapper[5109]: I1211 17:05:32.909110 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:32 crc kubenswrapper[5109]: I1211 17:05:32.955498 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:33 crc kubenswrapper[5109]: I1211 17:05:33.325297 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:33 crc kubenswrapper[5109]: I1211 17:05:33.370032 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-996cl"] Dec 11 17:05:35 crc kubenswrapper[5109]: I1211 17:05:35.272786 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-996cl" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="registry-server" containerID="cri-o://af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538" gracePeriod=2 Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.254215 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.280732 5109 generic.go:358] "Generic (PLEG): container finished" podID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerID="af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538" exitCode=0 Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.280831 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-996cl" event={"ID":"ad7a318c-68d9-48a9-bc3d-c7e99b64b186","Type":"ContainerDied","Data":"af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538"} Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.280881 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-996cl" event={"ID":"ad7a318c-68d9-48a9-bc3d-c7e99b64b186","Type":"ContainerDied","Data":"0e0617fe183cd2d058526738bd286209567449f17314daa28c62f7dc858aa364"} Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.280895 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-996cl" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.280900 5109 scope.go:117] "RemoveContainer" containerID="af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.300790 5109 scope.go:117] "RemoveContainer" containerID="23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.318894 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trbd5\" (UniqueName: \"kubernetes.io/projected/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-kube-api-access-trbd5\") pod \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.318963 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-utilities\") pod \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.319064 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-catalog-content\") pod \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\" (UID: \"ad7a318c-68d9-48a9-bc3d-c7e99b64b186\") " Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.322239 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-utilities" (OuterVolumeSpecName: "utilities") pod "ad7a318c-68d9-48a9-bc3d-c7e99b64b186" (UID: "ad7a318c-68d9-48a9-bc3d-c7e99b64b186"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.322363 5109 scope.go:117] "RemoveContainer" containerID="bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.326188 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-kube-api-access-trbd5" (OuterVolumeSpecName: "kube-api-access-trbd5") pod "ad7a318c-68d9-48a9-bc3d-c7e99b64b186" (UID: "ad7a318c-68d9-48a9-bc3d-c7e99b64b186"). InnerVolumeSpecName "kube-api-access-trbd5". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.363530 5109 scope.go:117] "RemoveContainer" containerID="af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538" Dec 11 17:05:36 crc kubenswrapper[5109]: E1211 17:05:36.364472 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538\": container with ID starting with af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538 not found: ID does not exist" containerID="af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.364520 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538"} err="failed to get container status \"af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538\": rpc error: code = NotFound desc = could not find container \"af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538\": container with ID starting with af9e0d943369c945afee6f795fd4219e5cde5025cd1c8e863370726e399cd538 not found: ID does not exist" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.364545 5109 scope.go:117] "RemoveContainer" containerID="23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695" Dec 11 17:05:36 crc kubenswrapper[5109]: E1211 17:05:36.364987 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695\": container with ID starting with 23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695 not found: ID does not exist" containerID="23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.365029 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695"} err="failed to get container status \"23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695\": rpc error: code = NotFound desc = could not find container \"23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695\": container with ID starting with 23d2b234ef4c1172fe796da16d8c175c1964c096799b41ae94244e19bce63695 not found: ID does not exist" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.365053 5109 scope.go:117] "RemoveContainer" containerID="bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5" Dec 11 17:05:36 crc kubenswrapper[5109]: E1211 17:05:36.365337 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5\": container with ID starting with bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5 not found: ID does not exist" containerID="bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.365366 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5"} err="failed to get container status \"bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5\": rpc error: code = NotFound desc = could not find container \"bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5\": container with ID starting with bcc3f85b2d47aba27e0528d6562e9ce015379b818cd800ab0b403fa0190d91a5 not found: ID does not exist" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.421297 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-trbd5\" (UniqueName: \"kubernetes.io/projected/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-kube-api-access-trbd5\") on node \"crc\" DevicePath \"\"" Dec 11 17:05:36 crc kubenswrapper[5109]: I1211 17:05:36.421368 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 17:05:37 crc kubenswrapper[5109]: I1211 17:05:37.511285 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ad7a318c-68d9-48a9-bc3d-c7e99b64b186" (UID: "ad7a318c-68d9-48a9-bc3d-c7e99b64b186"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:05:37 crc kubenswrapper[5109]: I1211 17:05:37.536248 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ad7a318c-68d9-48a9-bc3d-c7e99b64b186-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 17:05:37 crc kubenswrapper[5109]: I1211 17:05:37.822073 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-996cl"] Dec 11 17:05:37 crc kubenswrapper[5109]: I1211 17:05:37.832345 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-996cl"] Dec 11 17:05:38 crc kubenswrapper[5109]: I1211 17:05:38.910810 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" path="/var/lib/kubelet/pods/ad7a318c-68d9-48a9-bc3d-c7e99b64b186/volumes" Dec 11 17:06:09 crc kubenswrapper[5109]: I1211 17:06:09.568572 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz8n"] Dec 11 17:06:09 crc kubenswrapper[5109]: I1211 17:06:09.569432 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-czz8n" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="registry-server" containerID="cri-o://65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c" gracePeriod=30 Dec 11 17:06:09 crc kubenswrapper[5109]: I1211 17:06:09.912912 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.107295 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xw99k\" (UniqueName: \"kubernetes.io/projected/e1f282da-749c-42a0-aea4-0ec0587bc5fc-kube-api-access-xw99k\") pod \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.107404 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-utilities\") pod \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.107505 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-catalog-content\") pod \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\" (UID: \"e1f282da-749c-42a0-aea4-0ec0587bc5fc\") " Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.109815 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-utilities" (OuterVolumeSpecName: "utilities") pod "e1f282da-749c-42a0-aea4-0ec0587bc5fc" (UID: "e1f282da-749c-42a0-aea4-0ec0587bc5fc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.116778 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f282da-749c-42a0-aea4-0ec0587bc5fc-kube-api-access-xw99k" (OuterVolumeSpecName: "kube-api-access-xw99k") pod "e1f282da-749c-42a0-aea4-0ec0587bc5fc" (UID: "e1f282da-749c-42a0-aea4-0ec0587bc5fc"). InnerVolumeSpecName "kube-api-access-xw99k". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.132669 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1f282da-749c-42a0-aea4-0ec0587bc5fc" (UID: "e1f282da-749c-42a0-aea4-0ec0587bc5fc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.209086 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.209138 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-xw99k\" (UniqueName: \"kubernetes.io/projected/e1f282da-749c-42a0-aea4-0ec0587bc5fc-kube-api-access-xw99k\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.209162 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1f282da-749c-42a0-aea4-0ec0587bc5fc-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.504924 5109 generic.go:358] "Generic (PLEG): container finished" podID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerID="65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c" exitCode=0 Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.505038 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz8n" event={"ID":"e1f282da-749c-42a0-aea4-0ec0587bc5fc","Type":"ContainerDied","Data":"65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c"} Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.505075 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-czz8n" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.505104 5109 scope.go:117] "RemoveContainer" containerID="65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.505085 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-czz8n" event={"ID":"e1f282da-749c-42a0-aea4-0ec0587bc5fc","Type":"ContainerDied","Data":"a04effc515755b74e9db9059c2283fd4b61969fc02b4b87ddf1ffa1e8a7a7b96"} Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.536590 5109 scope.go:117] "RemoveContainer" containerID="2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.570275 5109 scope.go:117] "RemoveContainer" containerID="aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.581009 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz8n"] Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.592149 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-czz8n"] Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.594867 5109 scope.go:117] "RemoveContainer" containerID="65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c" Dec 11 17:06:10 crc kubenswrapper[5109]: E1211 17:06:10.595178 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c\": container with ID starting with 65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c not found: ID does not exist" containerID="65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.595212 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c"} err="failed to get container status \"65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c\": rpc error: code = NotFound desc = could not find container \"65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c\": container with ID starting with 65d73418838e8ca53ff43de43217f2ac5ff4077856badab8578f470b5c52072c not found: ID does not exist" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.595231 5109 scope.go:117] "RemoveContainer" containerID="2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319" Dec 11 17:06:10 crc kubenswrapper[5109]: E1211 17:06:10.595402 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319\": container with ID starting with 2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319 not found: ID does not exist" containerID="2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.595416 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319"} err="failed to get container status \"2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319\": rpc error: code = NotFound desc = could not find container \"2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319\": container with ID starting with 2998e6a433a2e377d6760bd00b27fddf2a102b737a80cd0c9a21080d6399e319 not found: ID does not exist" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.595429 5109 scope.go:117] "RemoveContainer" containerID="aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b" Dec 11 17:06:10 crc kubenswrapper[5109]: E1211 17:06:10.595578 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b\": container with ID starting with aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b not found: ID does not exist" containerID="aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.595591 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b"} err="failed to get container status \"aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b\": rpc error: code = NotFound desc = could not find container \"aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b\": container with ID starting with aa25204aa15176d6cf0dd0f3192b3eb297a6c0413f20b138caa40c0ec719d86b not found: ID does not exist" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.652973 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-bdrgv"] Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653597 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="extract-utilities" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653617 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="extract-utilities" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653650 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="extract-content" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653659 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="extract-content" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653674 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="extract-utilities" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653681 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="extract-utilities" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653694 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="registry-server" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653702 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="registry-server" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653717 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="registry-server" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653725 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="registry-server" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653758 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="extract-content" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653766 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="extract-content" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653878 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" containerName="registry-server" Dec 11 17:06:10 crc kubenswrapper[5109]: I1211 17:06:10.653895 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="ad7a318c-68d9-48a9-bc3d-c7e99b64b186" containerName="registry-server" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.744073 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.759117 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1f282da-749c-42a0-aea4-0ec0587bc5fc" path="/var/lib/kubelet/pods/e1f282da-749c-42a0-aea4-0ec0587bc5fc/volumes" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.760655 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-bdrgv"] Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808095 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808158 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-bound-sa-token\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808205 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/40ceacf7-9750-4610-9cc0-13105ad21d6e-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808242 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/40ceacf7-9750-4610-9cc0-13105ad21d6e-trusted-ca\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808264 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-registry-tls\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808290 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhn4q\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-kube-api-access-jhn4q\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808317 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/40ceacf7-9750-4610-9cc0-13105ad21d6e-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.808336 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/40ceacf7-9750-4610-9cc0-13105ad21d6e-registry-certificates\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.851396 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.909329 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-bound-sa-token\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.909664 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/40ceacf7-9750-4610-9cc0-13105ad21d6e-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.909868 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/40ceacf7-9750-4610-9cc0-13105ad21d6e-trusted-ca\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.911180 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-registry-tls\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.911140 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/40ceacf7-9750-4610-9cc0-13105ad21d6e-trusted-ca\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.912048 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jhn4q\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-kube-api-access-jhn4q\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.912166 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/40ceacf7-9750-4610-9cc0-13105ad21d6e-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.912246 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/40ceacf7-9750-4610-9cc0-13105ad21d6e-registry-certificates\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.912869 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/40ceacf7-9750-4610-9cc0-13105ad21d6e-ca-trust-extracted\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.913463 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/40ceacf7-9750-4610-9cc0-13105ad21d6e-registry-certificates\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.916015 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-registry-tls\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.928326 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-bound-sa-token\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.928690 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/40ceacf7-9750-4610-9cc0-13105ad21d6e-installation-pull-secrets\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:11 crc kubenswrapper[5109]: I1211 17:06:11.936756 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhn4q\" (UniqueName: \"kubernetes.io/projected/40ceacf7-9750-4610-9cc0-13105ad21d6e-kube-api-access-jhn4q\") pod \"image-registry-5d9d95bf5b-bdrgv\" (UID: \"40ceacf7-9750-4610-9cc0-13105ad21d6e\") " pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:12 crc kubenswrapper[5109]: I1211 17:06:12.064725 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:12 crc kubenswrapper[5109]: I1211 17:06:12.267758 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-5d9d95bf5b-bdrgv"] Dec 11 17:06:12 crc kubenswrapper[5109]: I1211 17:06:12.520039 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" event={"ID":"40ceacf7-9750-4610-9cc0-13105ad21d6e","Type":"ContainerStarted","Data":"f52733f323a3f9e08a3b653ebacbe9511fb7aef76d9bd0de641c5291380bbbe5"} Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.465717 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5"] Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.702419 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" event={"ID":"40ceacf7-9750-4610-9cc0-13105ad21d6e","Type":"ContainerStarted","Data":"613d6e16f64100f9ce7834021880183f08332da7083e93d42fbb41c6aa0393d2"} Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.702485 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5"] Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.702719 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.702799 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.706109 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.726407 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" podStartSLOduration=3.72638675 podStartE2EDuration="3.72638675s" podCreationTimestamp="2025-12-11 17:06:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 17:06:13.722327001 +0000 UTC m=+811.402018477" watchObservedRunningTime="2025-12-11 17:06:13.72638675 +0000 UTC m=+811.406078216" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.738399 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.738995 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.739124 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzlgm\" (UniqueName: \"kubernetes.io/projected/541f47de-3662-4e80-8a45-91ef221cb4e0-kube-api-access-lzlgm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.840894 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.840958 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-lzlgm\" (UniqueName: \"kubernetes.io/projected/541f47de-3662-4e80-8a45-91ef221cb4e0-kube-api-access-lzlgm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.841196 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.841465 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.841614 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:13 crc kubenswrapper[5109]: I1211 17:06:13.859972 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzlgm\" (UniqueName: \"kubernetes.io/projected/541f47de-3662-4e80-8a45-91ef221cb4e0-kube-api-access-lzlgm\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:14 crc kubenswrapper[5109]: I1211 17:06:14.021361 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:14 crc kubenswrapper[5109]: I1211 17:06:14.238166 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5"] Dec 11 17:06:14 crc kubenswrapper[5109]: I1211 17:06:14.534035 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" event={"ID":"541f47de-3662-4e80-8a45-91ef221cb4e0","Type":"ContainerStarted","Data":"830505ce2c72ce3cba466d5857eca00f1104b11d51de4d4e48f1b73c41be838b"} Dec 11 17:06:15 crc kubenswrapper[5109]: I1211 17:06:15.540190 5109 generic.go:358] "Generic (PLEG): container finished" podID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerID="a25dc0fea7b7de4706c5429d2f4e3f67b657e228a70775779a08f58bc042bae4" exitCode=0 Dec 11 17:06:15 crc kubenswrapper[5109]: I1211 17:06:15.540307 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" event={"ID":"541f47de-3662-4e80-8a45-91ef221cb4e0","Type":"ContainerDied","Data":"a25dc0fea7b7de4706c5429d2f4e3f67b657e228a70775779a08f58bc042bae4"} Dec 11 17:06:17 crc kubenswrapper[5109]: I1211 17:06:17.559712 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" event={"ID":"541f47de-3662-4e80-8a45-91ef221cb4e0","Type":"ContainerStarted","Data":"f56a80b737275281f6ecc7d957598c95cd5012569af30be53cedfd0115c3e490"} Dec 11 17:06:18 crc kubenswrapper[5109]: I1211 17:06:18.570664 5109 generic.go:358] "Generic (PLEG): container finished" podID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerID="f56a80b737275281f6ecc7d957598c95cd5012569af30be53cedfd0115c3e490" exitCode=0 Dec 11 17:06:18 crc kubenswrapper[5109]: I1211 17:06:18.570788 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" event={"ID":"541f47de-3662-4e80-8a45-91ef221cb4e0","Type":"ContainerDied","Data":"f56a80b737275281f6ecc7d957598c95cd5012569af30be53cedfd0115c3e490"} Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.043073 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx"] Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.050275 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.066365 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx"] Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.122637 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.122715 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.122877 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khcln\" (UniqueName: \"kubernetes.io/projected/c78662aa-cca8-42d5-b3d2-f5735c19c417-kube-api-access-khcln\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.224303 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.224361 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.224427 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-khcln\" (UniqueName: \"kubernetes.io/projected/c78662aa-cca8-42d5-b3d2-f5735c19c417-kube-api-access-khcln\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.224808 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-bundle\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.224878 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-util\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.260077 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-khcln\" (UniqueName: \"kubernetes.io/projected/c78662aa-cca8-42d5-b3d2-f5735c19c417-kube-api-access-khcln\") pod \"8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.379458 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.579457 5109 generic.go:358] "Generic (PLEG): container finished" podID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerID="be5f9c9610cbf7fb6092fb687d8b21721752274a853dc2197a38ac35c2c576d6" exitCode=0 Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.579696 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" event={"ID":"541f47de-3662-4e80-8a45-91ef221cb4e0","Type":"ContainerDied","Data":"be5f9c9610cbf7fb6092fb687d8b21721752274a853dc2197a38ac35c2c576d6"} Dec 11 17:06:19 crc kubenswrapper[5109]: I1211 17:06:19.591704 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx"] Dec 11 17:06:19 crc kubenswrapper[5109]: W1211 17:06:19.608155 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc78662aa_cca8_42d5_b3d2_f5735c19c417.slice/crio-44137cddb8169eb0a7705a5b95745ac070882701f34469193a1c908e618248e3 WatchSource:0}: Error finding container 44137cddb8169eb0a7705a5b95745ac070882701f34469193a1c908e618248e3: Status 404 returned error can't find the container with id 44137cddb8169eb0a7705a5b95745ac070882701f34469193a1c908e618248e3 Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.589530 5109 generic.go:358] "Generic (PLEG): container finished" podID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerID="37f56ba5bdf2c02df5a942847271b809df04c0f23b8bd4f6532d2a11cd5db0a4" exitCode=0 Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.589590 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" event={"ID":"c78662aa-cca8-42d5-b3d2-f5735c19c417","Type":"ContainerDied","Data":"37f56ba5bdf2c02df5a942847271b809df04c0f23b8bd4f6532d2a11cd5db0a4"} Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.590075 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" event={"ID":"c78662aa-cca8-42d5-b3d2-f5735c19c417","Type":"ContainerStarted","Data":"44137cddb8169eb0a7705a5b95745ac070882701f34469193a1c908e618248e3"} Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.881765 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.946668 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-bundle\") pod \"541f47de-3662-4e80-8a45-91ef221cb4e0\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.946722 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-util\") pod \"541f47de-3662-4e80-8a45-91ef221cb4e0\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.946902 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzlgm\" (UniqueName: \"kubernetes.io/projected/541f47de-3662-4e80-8a45-91ef221cb4e0-kube-api-access-lzlgm\") pod \"541f47de-3662-4e80-8a45-91ef221cb4e0\" (UID: \"541f47de-3662-4e80-8a45-91ef221cb4e0\") " Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.953033 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/541f47de-3662-4e80-8a45-91ef221cb4e0-kube-api-access-lzlgm" (OuterVolumeSpecName: "kube-api-access-lzlgm") pod "541f47de-3662-4e80-8a45-91ef221cb4e0" (UID: "541f47de-3662-4e80-8a45-91ef221cb4e0"). InnerVolumeSpecName "kube-api-access-lzlgm". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.956817 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-util" (OuterVolumeSpecName: "util") pod "541f47de-3662-4e80-8a45-91ef221cb4e0" (UID: "541f47de-3662-4e80-8a45-91ef221cb4e0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:20 crc kubenswrapper[5109]: I1211 17:06:20.959637 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-bundle" (OuterVolumeSpecName: "bundle") pod "541f47de-3662-4e80-8a45-91ef221cb4e0" (UID: "541f47de-3662-4e80-8a45-91ef221cb4e0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.048729 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-lzlgm\" (UniqueName: \"kubernetes.io/projected/541f47de-3662-4e80-8a45-91ef221cb4e0-kube-api-access-lzlgm\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.048781 5109 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.048791 5109 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/541f47de-3662-4e80-8a45-91ef221cb4e0-util\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.601765 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" event={"ID":"541f47de-3662-4e80-8a45-91ef221cb4e0","Type":"ContainerDied","Data":"830505ce2c72ce3cba466d5857eca00f1104b11d51de4d4e48f1b73c41be838b"} Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.602136 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="830505ce2c72ce3cba466d5857eca00f1104b11d51de4d4e48f1b73c41be838b" Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.601848 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c921046gx5" Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.603399 5109 generic.go:358] "Generic (PLEG): container finished" podID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerID="ba8f7b2a383c3e2aafe8b0055593b407c4e2636940d23ca0c3d8647a0571ae93" exitCode=0 Dec 11 17:06:21 crc kubenswrapper[5109]: I1211 17:06:21.603489 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" event={"ID":"c78662aa-cca8-42d5-b3d2-f5735c19c417","Type":"ContainerDied","Data":"ba8f7b2a383c3e2aafe8b0055593b407c4e2636940d23ca0c3d8647a0571ae93"} Dec 11 17:06:22 crc kubenswrapper[5109]: I1211 17:06:22.623165 5109 generic.go:358] "Generic (PLEG): container finished" podID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerID="30c8e04fa221c523641bbfc1366e6166457ab251d9702fb2fa4c704f57d06f33" exitCode=0 Dec 11 17:06:22 crc kubenswrapper[5109]: I1211 17:06:22.624119 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" event={"ID":"c78662aa-cca8-42d5-b3d2-f5735c19c417","Type":"ContainerDied","Data":"30c8e04fa221c523641bbfc1366e6166457ab251d9702fb2fa4c704f57d06f33"} Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.663215 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst"] Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.663893 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerName="extract" Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.663908 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerName="extract" Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.663936 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerName="pull" Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.663944 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerName="pull" Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.663976 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerName="util" Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.663984 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerName="util" Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.664091 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="541f47de-3662-4e80-8a45-91ef221cb4e0" containerName="extract" Dec 11 17:06:23 crc kubenswrapper[5109]: I1211 17:06:23.914029 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.001033 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-khcln\" (UniqueName: \"kubernetes.io/projected/c78662aa-cca8-42d5-b3d2-f5735c19c417-kube-api-access-khcln\") pod \"c78662aa-cca8-42d5-b3d2-f5735c19c417\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.001160 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-util\") pod \"c78662aa-cca8-42d5-b3d2-f5735c19c417\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.001248 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-bundle\") pod \"c78662aa-cca8-42d5-b3d2-f5735c19c417\" (UID: \"c78662aa-cca8-42d5-b3d2-f5735c19c417\") " Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.002246 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-bundle" (OuterVolumeSpecName: "bundle") pod "c78662aa-cca8-42d5-b3d2-f5735c19c417" (UID: "c78662aa-cca8-42d5-b3d2-f5735c19c417"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.012486 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c78662aa-cca8-42d5-b3d2-f5735c19c417-kube-api-access-khcln" (OuterVolumeSpecName: "kube-api-access-khcln") pod "c78662aa-cca8-42d5-b3d2-f5735c19c417" (UID: "c78662aa-cca8-42d5-b3d2-f5735c19c417"). InnerVolumeSpecName "kube-api-access-khcln". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.020330 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-util" (OuterVolumeSpecName: "util") pod "c78662aa-cca8-42d5-b3d2-f5735c19c417" (UID: "c78662aa-cca8-42d5-b3d2-f5735c19c417"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.102318 5109 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-util\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.102354 5109 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c78662aa-cca8-42d5-b3d2-f5735c19c417-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.102363 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-khcln\" (UniqueName: \"kubernetes.io/projected/c78662aa-cca8-42d5-b3d2-f5735c19c417-kube-api-access-khcln\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.150355 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst"] Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.150436 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mpfqq"] Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.150594 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.151198 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerName="pull" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.151239 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerName="pull" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.151260 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerName="util" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.151269 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerName="util" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.151279 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerName="extract" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.151286 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerName="extract" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.151401 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="c78662aa-cca8-42d5-b3d2-f5735c19c417" containerName="extract" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.176212 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mpfqq"] Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.176389 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.203808 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.203947 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5rmq\" (UniqueName: \"kubernetes.io/projected/cdf5d610-1865-462b-91b3-6cca426020ce-kube-api-access-d5rmq\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.203982 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.305395 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-d5rmq\" (UniqueName: \"kubernetes.io/projected/cdf5d610-1865-462b-91b3-6cca426020ce-kube-api-access-d5rmq\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.305700 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.305767 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-catalog-content\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.305816 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dch5m\" (UniqueName: \"kubernetes.io/projected/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-kube-api-access-dch5m\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.305862 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.305996 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-utilities\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.306258 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-bundle\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.306324 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-util\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.343878 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5rmq\" (UniqueName: \"kubernetes.io/projected/cdf5d610-1865-462b-91b3-6cca426020ce-kube-api-access-d5rmq\") pod \"1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.407311 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-utilities\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.407402 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-catalog-content\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.407428 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-dch5m\" (UniqueName: \"kubernetes.io/projected/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-kube-api-access-dch5m\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.407930 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-utilities\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.407956 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-catalog-content\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.424585 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-dch5m\" (UniqueName: \"kubernetes.io/projected/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-kube-api-access-dch5m\") pod \"certified-operators-mpfqq\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.463294 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.487917 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.649774 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" event={"ID":"c78662aa-cca8-42d5-b3d2-f5735c19c417","Type":"ContainerDied","Data":"44137cddb8169eb0a7705a5b95745ac070882701f34469193a1c908e618248e3"} Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.649827 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44137cddb8169eb0a7705a5b95745ac070882701f34469193a1c908e618248e3" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.649852 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8ed862a309935d5a1c8012df79b93f7fb46e029d4689f7f6ddcb9e7f5entkcx" Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.676604 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst"] Dec 11 17:06:24 crc kubenswrapper[5109]: I1211 17:06:24.948599 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mpfqq"] Dec 11 17:06:25 crc kubenswrapper[5109]: I1211 17:06:25.655748 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mpfqq" event={"ID":"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291","Type":"ContainerStarted","Data":"87618a61602e4e32a3487b70bcb344ef05f1f9c41a22c081d6206a86017a9c58"} Dec 11 17:06:25 crc kubenswrapper[5109]: I1211 17:06:25.657228 5109 generic.go:358] "Generic (PLEG): container finished" podID="cdf5d610-1865-462b-91b3-6cca426020ce" containerID="98feee118aeef5030fb9d658389441dceccd6805e5a9e5aeb9fa1fd7d13ee81d" exitCode=0 Dec 11 17:06:25 crc kubenswrapper[5109]: I1211 17:06:25.657287 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" event={"ID":"cdf5d610-1865-462b-91b3-6cca426020ce","Type":"ContainerDied","Data":"98feee118aeef5030fb9d658389441dceccd6805e5a9e5aeb9fa1fd7d13ee81d"} Dec 11 17:06:25 crc kubenswrapper[5109]: I1211 17:06:25.657302 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" event={"ID":"cdf5d610-1865-462b-91b3-6cca426020ce","Type":"ContainerStarted","Data":"8dcd9bd59d8234753aaed77f5f52ece02eb28c27c4877cd3422df87f5c59a918"} Dec 11 17:06:26 crc kubenswrapper[5109]: I1211 17:06:26.663954 5109 generic.go:358] "Generic (PLEG): container finished" podID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerID="44e234f6b7a03eb22d2656aabe010b94846d783cb0bb0b707be71d814c6786a3" exitCode=0 Dec 11 17:06:26 crc kubenswrapper[5109]: I1211 17:06:26.664067 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mpfqq" event={"ID":"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291","Type":"ContainerDied","Data":"44e234f6b7a03eb22d2656aabe010b94846d783cb0bb0b707be71d814c6786a3"} Dec 11 17:06:28 crc kubenswrapper[5109]: I1211 17:06:28.679222 5109 generic.go:358] "Generic (PLEG): container finished" podID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerID="cfb5594ed0f858b5c0ac95a5fc1e5dbc514f990e93ab5c72b97fe35255ebb685" exitCode=0 Dec 11 17:06:28 crc kubenswrapper[5109]: I1211 17:06:28.679280 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mpfqq" event={"ID":"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291","Type":"ContainerDied","Data":"cfb5594ed0f858b5c0ac95a5fc1e5dbc514f990e93ab5c72b97fe35255ebb685"} Dec 11 17:06:30 crc kubenswrapper[5109]: I1211 17:06:30.710041 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" event={"ID":"cdf5d610-1865-462b-91b3-6cca426020ce","Type":"ContainerStarted","Data":"3a6f1c3172745b9708048cf806cd09fdb64f9702625b4cb87d4608b0461ed51c"} Dec 11 17:06:31 crc kubenswrapper[5109]: I1211 17:06:31.719012 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mpfqq" event={"ID":"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291","Type":"ContainerStarted","Data":"33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4"} Dec 11 17:06:31 crc kubenswrapper[5109]: I1211 17:06:31.720341 5109 generic.go:358] "Generic (PLEG): container finished" podID="cdf5d610-1865-462b-91b3-6cca426020ce" containerID="3a6f1c3172745b9708048cf806cd09fdb64f9702625b4cb87d4608b0461ed51c" exitCode=0 Dec 11 17:06:31 crc kubenswrapper[5109]: I1211 17:06:31.720402 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" event={"ID":"cdf5d610-1865-462b-91b3-6cca426020ce","Type":"ContainerDied","Data":"3a6f1c3172745b9708048cf806cd09fdb64f9702625b4cb87d4608b0461ed51c"} Dec 11 17:06:31 crc kubenswrapper[5109]: I1211 17:06:31.739544 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mpfqq" podStartSLOduration=7.806178335 podStartE2EDuration="8.73952294s" podCreationTimestamp="2025-12-11 17:06:23 +0000 UTC" firstStartedPulling="2025-12-11 17:06:26.664937544 +0000 UTC m=+824.344629010" lastFinishedPulling="2025-12-11 17:06:27.598282149 +0000 UTC m=+825.277973615" observedRunningTime="2025-12-11 17:06:31.735866592 +0000 UTC m=+829.415558058" watchObservedRunningTime="2025-12-11 17:06:31.73952294 +0000 UTC m=+829.419214406" Dec 11 17:06:32 crc kubenswrapper[5109]: I1211 17:06:32.728625 5109 generic.go:358] "Generic (PLEG): container finished" podID="cdf5d610-1865-462b-91b3-6cca426020ce" containerID="a2c950a9a12611a366468daa41ccc228541684afa2b58a75b3ce82bad377b6ae" exitCode=0 Dec 11 17:06:32 crc kubenswrapper[5109]: I1211 17:06:32.728720 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" event={"ID":"cdf5d610-1865-462b-91b3-6cca426020ce","Type":"ContainerDied","Data":"a2c950a9a12611a366468daa41ccc228541684afa2b58a75b3ce82bad377b6ae"} Dec 11 17:06:33 crc kubenswrapper[5109]: I1211 17:06:33.447823 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-c46955c7-5p87m"] Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.007630 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-c46955c7-5p87m"] Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.007827 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.012622 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elastic-operator-service-cert\"" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.012980 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elastic-operator-dockercfg-nc499\"" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.013146 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"kube-root-ca.crt\"" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.013238 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"openshift-service-ca.crt\"" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.124975 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/907ddecb-77b2-496b-88c3-bc9bff555b95-webhook-cert\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.125102 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/907ddecb-77b2-496b-88c3-bc9bff555b95-apiservice-cert\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.125185 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhvnn\" (UniqueName: \"kubernetes.io/projected/907ddecb-77b2-496b-88c3-bc9bff555b95-kube-api-access-qhvnn\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.228339 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/907ddecb-77b2-496b-88c3-bc9bff555b95-webhook-cert\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.228396 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/907ddecb-77b2-496b-88c3-bc9bff555b95-apiservice-cert\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.228432 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qhvnn\" (UniqueName: \"kubernetes.io/projected/907ddecb-77b2-496b-88c3-bc9bff555b95-kube-api-access-qhvnn\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.235988 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/907ddecb-77b2-496b-88c3-bc9bff555b95-webhook-cert\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.236038 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/907ddecb-77b2-496b-88c3-bc9bff555b95-apiservice-cert\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.246933 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhvnn\" (UniqueName: \"kubernetes.io/projected/907ddecb-77b2-496b-88c3-bc9bff555b95-kube-api-access-qhvnn\") pod \"elastic-operator-c46955c7-5p87m\" (UID: \"907ddecb-77b2-496b-88c3-bc9bff555b95\") " pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.280438 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.323697 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-c46955c7-5p87m" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.328779 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5rmq\" (UniqueName: \"kubernetes.io/projected/cdf5d610-1865-462b-91b3-6cca426020ce-kube-api-access-d5rmq\") pod \"cdf5d610-1865-462b-91b3-6cca426020ce\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.328813 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-bundle\") pod \"cdf5d610-1865-462b-91b3-6cca426020ce\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.328887 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-util\") pod \"cdf5d610-1865-462b-91b3-6cca426020ce\" (UID: \"cdf5d610-1865-462b-91b3-6cca426020ce\") " Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.338108 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-util" (OuterVolumeSpecName: "util") pod "cdf5d610-1865-462b-91b3-6cca426020ce" (UID: "cdf5d610-1865-462b-91b3-6cca426020ce"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.339646 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-bundle" (OuterVolumeSpecName: "bundle") pod "cdf5d610-1865-462b-91b3-6cca426020ce" (UID: "cdf5d610-1865-462b-91b3-6cca426020ce"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.343023 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdf5d610-1865-462b-91b3-6cca426020ce-kube-api-access-d5rmq" (OuterVolumeSpecName: "kube-api-access-d5rmq") pod "cdf5d610-1865-462b-91b3-6cca426020ce" (UID: "cdf5d610-1865-462b-91b3-6cca426020ce"). InnerVolumeSpecName "kube-api-access-d5rmq". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.430376 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-d5rmq\" (UniqueName: \"kubernetes.io/projected/cdf5d610-1865-462b-91b3-6cca426020ce-kube-api-access-d5rmq\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.430724 5109 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.430753 5109 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/cdf5d610-1865-462b-91b3-6cca426020ce-util\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.488960 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.489012 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.533647 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-c46955c7-5p87m"] Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.538726 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.540248 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-5d9d95bf5b-bdrgv" Dec 11 17:06:34 crc kubenswrapper[5109]: W1211 17:06:34.548066 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod907ddecb_77b2_496b_88c3_bc9bff555b95.slice/crio-459c8dc48b9c7ece6e741e8424ee5af23032bf0d61352870f386e2e7f67da6a2 WatchSource:0}: Error finding container 459c8dc48b9c7ece6e741e8424ee5af23032bf0d61352870f386e2e7f67da6a2: Status 404 returned error can't find the container with id 459c8dc48b9c7ece6e741e8424ee5af23032bf0d61352870f386e2e7f67da6a2 Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.625405 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-nkxtk"] Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.741582 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.741598 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1f59f640c8a0eb1a7b0f26c81382bbdde784d03eb439a940bb8da3931apppst" event={"ID":"cdf5d610-1865-462b-91b3-6cca426020ce","Type":"ContainerDied","Data":"8dcd9bd59d8234753aaed77f5f52ece02eb28c27c4877cd3422df87f5c59a918"} Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.741634 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8dcd9bd59d8234753aaed77f5f52ece02eb28c27c4877cd3422df87f5c59a918" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.743491 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-c46955c7-5p87m" event={"ID":"907ddecb-77b2-496b-88c3-bc9bff555b95","Type":"ContainerStarted","Data":"459c8dc48b9c7ece6e741e8424ee5af23032bf0d61352870f386e2e7f67da6a2"} Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.888367 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-86648f486b-h4fwb"] Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.888919 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="cdf5d610-1865-462b-91b3-6cca426020ce" containerName="util" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.888935 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf5d610-1865-462b-91b3-6cca426020ce" containerName="util" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.888946 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="cdf5d610-1865-462b-91b3-6cca426020ce" containerName="pull" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.888952 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf5d610-1865-462b-91b3-6cca426020ce" containerName="pull" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.888962 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="cdf5d610-1865-462b-91b3-6cca426020ce" containerName="extract" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.888971 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="cdf5d610-1865-462b-91b3-6cca426020ce" containerName="extract" Dec 11 17:06:34 crc kubenswrapper[5109]: I1211 17:06:34.889064 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="cdf5d610-1865-462b-91b3-6cca426020ce" containerName="extract" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.295649 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.301467 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operators\"/\"kube-root-ca.crt\"" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.301572 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operators\"/\"openshift-service-ca.crt\"" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.302037 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-dockercfg-f6mvz\"" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.308203 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-86648f486b-h4fwb"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.308268 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.343517 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwpqn\" (UniqueName: \"kubernetes.io/projected/0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77-kube-api-access-fwpqn\") pod \"obo-prometheus-operator-86648f486b-h4fwb\" (UID: \"0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77\") " pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.445299 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-fwpqn\" (UniqueName: \"kubernetes.io/projected/0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77-kube-api-access-fwpqn\") pod \"obo-prometheus-operator-86648f486b-h4fwb\" (UID: \"0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77\") " pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.481760 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwpqn\" (UniqueName: \"kubernetes.io/projected/0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77-kube-api-access-fwpqn\") pod \"obo-prometheus-operator-86648f486b-h4fwb\" (UID: \"0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77\") " pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.629145 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.959047 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.959346 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.959309 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.964492 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-dockercfg-vfs2t\"" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.966361 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"obo-prometheus-operator-admission-webhook-service-cert\"" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.980504 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.980557 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-78c97476f4-8rw8j"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.980755 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.985673 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-78c97476f4-8rw8j"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.985701 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-68bdb49cbf-sqjbr"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.986777 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.989599 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"observability-operator-tls\"" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.989681 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"observability-operator-sa-dockercfg-dx2jq\"" Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.990697 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-68bdb49cbf-sqjbr"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.990723 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-86648f486b-h4fwb"] Dec 11 17:06:35 crc kubenswrapper[5109]: I1211 17:06:35.990840 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.003500 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operators\"/\"perses-operator-dockercfg-qlbsq\"" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.052817 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/042ee66e-ed7c-44e5-aa2d-bcc5bea13707-observability-operator-tls\") pod \"observability-operator-78c97476f4-8rw8j\" (UID: \"042ee66e-ed7c-44e5-aa2d-bcc5bea13707\") " pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.052895 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/116f2f7e-6d09-43af-9754-962ae5909372-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc\" (UID: \"116f2f7e-6d09-43af-9754-962ae5909372\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.052996 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvppw\" (UniqueName: \"kubernetes.io/projected/042ee66e-ed7c-44e5-aa2d-bcc5bea13707-kube-api-access-cvppw\") pod \"observability-operator-78c97476f4-8rw8j\" (UID: \"042ee66e-ed7c-44e5-aa2d-bcc5bea13707\") " pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.053067 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5432d60-106c-4c32-b8f4-e4064b8a71c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt\" (UID: \"c5432d60-106c-4c32-b8f4-e4064b8a71c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.053151 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/116f2f7e-6d09-43af-9754-962ae5909372-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc\" (UID: \"116f2f7e-6d09-43af-9754-962ae5909372\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.053258 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd35d77f-5c0e-4a54-9a68-31e729d364bf-openshift-service-ca\") pod \"perses-operator-68bdb49cbf-sqjbr\" (UID: \"dd35d77f-5c0e-4a54-9a68-31e729d364bf\") " pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.053293 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrdcc\" (UniqueName: \"kubernetes.io/projected/dd35d77f-5c0e-4a54-9a68-31e729d364bf-kube-api-access-xrdcc\") pod \"perses-operator-68bdb49cbf-sqjbr\" (UID: \"dd35d77f-5c0e-4a54-9a68-31e729d364bf\") " pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.053321 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5432d60-106c-4c32-b8f4-e4064b8a71c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt\" (UID: \"c5432d60-106c-4c32-b8f4-e4064b8a71c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154338 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xrdcc\" (UniqueName: \"kubernetes.io/projected/dd35d77f-5c0e-4a54-9a68-31e729d364bf-kube-api-access-xrdcc\") pod \"perses-operator-68bdb49cbf-sqjbr\" (UID: \"dd35d77f-5c0e-4a54-9a68-31e729d364bf\") " pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154398 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5432d60-106c-4c32-b8f4-e4064b8a71c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt\" (UID: \"c5432d60-106c-4c32-b8f4-e4064b8a71c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154443 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/042ee66e-ed7c-44e5-aa2d-bcc5bea13707-observability-operator-tls\") pod \"observability-operator-78c97476f4-8rw8j\" (UID: \"042ee66e-ed7c-44e5-aa2d-bcc5bea13707\") " pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154478 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/116f2f7e-6d09-43af-9754-962ae5909372-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc\" (UID: \"116f2f7e-6d09-43af-9754-962ae5909372\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154505 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cvppw\" (UniqueName: \"kubernetes.io/projected/042ee66e-ed7c-44e5-aa2d-bcc5bea13707-kube-api-access-cvppw\") pod \"observability-operator-78c97476f4-8rw8j\" (UID: \"042ee66e-ed7c-44e5-aa2d-bcc5bea13707\") " pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154536 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5432d60-106c-4c32-b8f4-e4064b8a71c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt\" (UID: \"c5432d60-106c-4c32-b8f4-e4064b8a71c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154583 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/116f2f7e-6d09-43af-9754-962ae5909372-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc\" (UID: \"116f2f7e-6d09-43af-9754-962ae5909372\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.154633 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd35d77f-5c0e-4a54-9a68-31e729d364bf-openshift-service-ca\") pod \"perses-operator-68bdb49cbf-sqjbr\" (UID: \"dd35d77f-5c0e-4a54-9a68-31e729d364bf\") " pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.156804 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/dd35d77f-5c0e-4a54-9a68-31e729d364bf-openshift-service-ca\") pod \"perses-operator-68bdb49cbf-sqjbr\" (UID: \"dd35d77f-5c0e-4a54-9a68-31e729d364bf\") " pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.159828 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/042ee66e-ed7c-44e5-aa2d-bcc5bea13707-observability-operator-tls\") pod \"observability-operator-78c97476f4-8rw8j\" (UID: \"042ee66e-ed7c-44e5-aa2d-bcc5bea13707\") " pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.159970 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c5432d60-106c-4c32-b8f4-e4064b8a71c8-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt\" (UID: \"c5432d60-106c-4c32-b8f4-e4064b8a71c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.161494 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/116f2f7e-6d09-43af-9754-962ae5909372-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc\" (UID: \"116f2f7e-6d09-43af-9754-962ae5909372\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.162458 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c5432d60-106c-4c32-b8f4-e4064b8a71c8-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt\" (UID: \"c5432d60-106c-4c32-b8f4-e4064b8a71c8\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.165289 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/116f2f7e-6d09-43af-9754-962ae5909372-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc\" (UID: \"116f2f7e-6d09-43af-9754-962ae5909372\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.174336 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvppw\" (UniqueName: \"kubernetes.io/projected/042ee66e-ed7c-44e5-aa2d-bcc5bea13707-kube-api-access-cvppw\") pod \"observability-operator-78c97476f4-8rw8j\" (UID: \"042ee66e-ed7c-44e5-aa2d-bcc5bea13707\") " pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.178058 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrdcc\" (UniqueName: \"kubernetes.io/projected/dd35d77f-5c0e-4a54-9a68-31e729d364bf-kube-api-access-xrdcc\") pod \"perses-operator-68bdb49cbf-sqjbr\" (UID: \"dd35d77f-5c0e-4a54-9a68-31e729d364bf\") " pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.279124 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.302625 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.317561 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.326918 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.771219 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" event={"ID":"0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77","Type":"ContainerStarted","Data":"ae0c6066f9bd729869ff8fad5ed268b6bcc8352551adcf068236c2ad981f0236"} Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.908785 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc"] Dec 11 17:06:36 crc kubenswrapper[5109]: W1211 17:06:36.917891 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod116f2f7e_6d09_43af_9754_962ae5909372.slice/crio-e67300280142e3d658d85779f768d10472f3664bc5119b34004ee86b79963511 WatchSource:0}: Error finding container e67300280142e3d658d85779f768d10472f3664bc5119b34004ee86b79963511: Status 404 returned error can't find the container with id e67300280142e3d658d85779f768d10472f3664bc5119b34004ee86b79963511 Dec 11 17:06:36 crc kubenswrapper[5109]: I1211 17:06:36.925242 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt"] Dec 11 17:06:36 crc kubenswrapper[5109]: W1211 17:06:36.939918 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc5432d60_106c_4c32_b8f4_e4064b8a71c8.slice/crio-8c9afadb156b3388bdf0d9152e320c0b4a64cae89eb527643f0d3a82908478ef WatchSource:0}: Error finding container 8c9afadb156b3388bdf0d9152e320c0b4a64cae89eb527643f0d3a82908478ef: Status 404 returned error can't find the container with id 8c9afadb156b3388bdf0d9152e320c0b4a64cae89eb527643f0d3a82908478ef Dec 11 17:06:37 crc kubenswrapper[5109]: I1211 17:06:37.043223 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-68bdb49cbf-sqjbr"] Dec 11 17:06:37 crc kubenswrapper[5109]: I1211 17:06:37.053454 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-78c97476f4-8rw8j"] Dec 11 17:06:37 crc kubenswrapper[5109]: I1211 17:06:37.856582 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" event={"ID":"116f2f7e-6d09-43af-9754-962ae5909372","Type":"ContainerStarted","Data":"e67300280142e3d658d85779f768d10472f3664bc5119b34004ee86b79963511"} Dec 11 17:06:37 crc kubenswrapper[5109]: I1211 17:06:37.863882 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" event={"ID":"c5432d60-106c-4c32-b8f4-e4064b8a71c8","Type":"ContainerStarted","Data":"8c9afadb156b3388bdf0d9152e320c0b4a64cae89eb527643f0d3a82908478ef"} Dec 11 17:06:39 crc kubenswrapper[5109]: I1211 17:06:39.883054 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-78c97476f4-8rw8j" event={"ID":"042ee66e-ed7c-44e5-aa2d-bcc5bea13707","Type":"ContainerStarted","Data":"1a05041d8b34878cff35cdde5d1d840a308076389cfc8a6f2612e52da9ef2bde"} Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.000593 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xc448"] Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.466335 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.466831 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xc448"] Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.522096 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-utilities\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.522179 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-catalog-content\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.522227 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2bws\" (UniqueName: \"kubernetes.io/projected/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-kube-api-access-n2bws\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.627009 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-utilities\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.627098 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-catalog-content\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.627169 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-n2bws\" (UniqueName: \"kubernetes.io/projected/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-kube-api-access-n2bws\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.627927 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-utilities\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.628158 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-catalog-content\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.669117 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2bws\" (UniqueName: \"kubernetes.io/projected/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-kube-api-access-n2bws\") pod \"community-operators-xc448\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: W1211 17:06:40.782304 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd35d77f_5c0e_4a54_9a68_31e729d364bf.slice/crio-4dd30a2cf3e7793468bcadd382b02f315c114c544b070b0628e680e599ebf8ca WatchSource:0}: Error finding container 4dd30a2cf3e7793468bcadd382b02f315c114c544b070b0628e680e599ebf8ca: Status 404 returned error can't find the container with id 4dd30a2cf3e7793468bcadd382b02f315c114c544b070b0628e680e599ebf8ca Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.788396 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:06:40 crc kubenswrapper[5109]: I1211 17:06:40.891037 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" event={"ID":"dd35d77f-5c0e-4a54-9a68-31e729d364bf","Type":"ContainerStarted","Data":"4dd30a2cf3e7793468bcadd382b02f315c114c544b070b0628e680e599ebf8ca"} Dec 11 17:06:41 crc kubenswrapper[5109]: I1211 17:06:41.234913 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xc448"] Dec 11 17:06:41 crc kubenswrapper[5109]: I1211 17:06:41.900897 5109 generic.go:358] "Generic (PLEG): container finished" podID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerID="522c2492c160a787504995fa3f8058b169438653e309eb226d61b9ec164f9a0a" exitCode=0 Dec 11 17:06:41 crc kubenswrapper[5109]: I1211 17:06:41.901054 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xc448" event={"ID":"ee0bc95f-e0a3-4ef8-aa90-71475f70929d","Type":"ContainerDied","Data":"522c2492c160a787504995fa3f8058b169438653e309eb226d61b9ec164f9a0a"} Dec 11 17:06:41 crc kubenswrapper[5109]: I1211 17:06:41.901087 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xc448" event={"ID":"ee0bc95f-e0a3-4ef8-aa90-71475f70929d","Type":"ContainerStarted","Data":"3efa8c5c650625d5cc8fd088a8d3c0d5f388de99b8b29d6849c0be33936a43e3"} Dec 11 17:06:41 crc kubenswrapper[5109]: I1211 17:06:41.903773 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-c46955c7-5p87m" event={"ID":"907ddecb-77b2-496b-88c3-bc9bff555b95","Type":"ContainerStarted","Data":"c031d991c6b23d4ea8d37876aff0d1aee6521bcb9305343a1e1bdbbffc5a2e16"} Dec 11 17:06:41 crc kubenswrapper[5109]: I1211 17:06:41.945994 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-c46955c7-5p87m" podStartSLOduration=2.56394477 podStartE2EDuration="8.945973996s" podCreationTimestamp="2025-12-11 17:06:33 +0000 UTC" firstStartedPulling="2025-12-11 17:06:34.551913232 +0000 UTC m=+832.231604698" lastFinishedPulling="2025-12-11 17:06:40.933942458 +0000 UTC m=+838.613633924" observedRunningTime="2025-12-11 17:06:41.945392891 +0000 UTC m=+839.625084377" watchObservedRunningTime="2025-12-11 17:06:41.945973996 +0000 UTC m=+839.625665462" Dec 11 17:06:44 crc kubenswrapper[5109]: I1211 17:06:44.786763 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.817087 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz"] Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.869443 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz"] Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.869598 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.877592 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"openshift-service-ca.crt\"" Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.878082 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager-operator\"/\"cert-manager-operator-controller-manager-dockercfg-t89tb\"" Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.878238 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager-operator\"/\"kube-root-ca.crt\"" Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.923645 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0e184e6-698a-40cc-ba2d-62e5e16b2980-tmp\") pod \"cert-manager-operator-controller-manager-64c74584c4-hdjzz\" (UID: \"c0e184e6-698a-40cc-ba2d-62e5e16b2980\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:46 crc kubenswrapper[5109]: I1211 17:06:46.923692 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhs2m\" (UniqueName: \"kubernetes.io/projected/c0e184e6-698a-40cc-ba2d-62e5e16b2980-kube-api-access-zhs2m\") pod \"cert-manager-operator-controller-manager-64c74584c4-hdjzz\" (UID: \"c0e184e6-698a-40cc-ba2d-62e5e16b2980\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:47 crc kubenswrapper[5109]: I1211 17:06:47.025815 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0e184e6-698a-40cc-ba2d-62e5e16b2980-tmp\") pod \"cert-manager-operator-controller-manager-64c74584c4-hdjzz\" (UID: \"c0e184e6-698a-40cc-ba2d-62e5e16b2980\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:47 crc kubenswrapper[5109]: I1211 17:06:47.025851 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zhs2m\" (UniqueName: \"kubernetes.io/projected/c0e184e6-698a-40cc-ba2d-62e5e16b2980-kube-api-access-zhs2m\") pod \"cert-manager-operator-controller-manager-64c74584c4-hdjzz\" (UID: \"c0e184e6-698a-40cc-ba2d-62e5e16b2980\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:47 crc kubenswrapper[5109]: I1211 17:06:47.026500 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/c0e184e6-698a-40cc-ba2d-62e5e16b2980-tmp\") pod \"cert-manager-operator-controller-manager-64c74584c4-hdjzz\" (UID: \"c0e184e6-698a-40cc-ba2d-62e5e16b2980\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:47 crc kubenswrapper[5109]: I1211 17:06:47.049890 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhs2m\" (UniqueName: \"kubernetes.io/projected/c0e184e6-698a-40cc-ba2d-62e5e16b2980-kube-api-access-zhs2m\") pod \"cert-manager-operator-controller-manager-64c74584c4-hdjzz\" (UID: \"c0e184e6-698a-40cc-ba2d-62e5e16b2980\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:47 crc kubenswrapper[5109]: I1211 17:06:47.196651 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" Dec 11 17:06:49 crc kubenswrapper[5109]: I1211 17:06:49.394092 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mpfqq"] Dec 11 17:06:49 crc kubenswrapper[5109]: I1211 17:06:49.395609 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mpfqq" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="registry-server" containerID="cri-o://33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4" gracePeriod=2 Dec 11 17:06:49 crc kubenswrapper[5109]: I1211 17:06:49.955240 5109 generic.go:358] "Generic (PLEG): container finished" podID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerID="33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4" exitCode=0 Dec 11 17:06:49 crc kubenswrapper[5109]: I1211 17:06:49.955301 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mpfqq" event={"ID":"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291","Type":"ContainerDied","Data":"33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4"} Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.202937 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.455980 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.456159 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.458711 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-default-es-config\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.458833 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-internal-users\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.458979 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-dockercfg-s8s2d\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.459049 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-default-es-transport-certs\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.459145 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"elasticsearch-es-unicast-hosts\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.461419 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-xpack-file-realm\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.462133 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"elasticsearch-es-scripts\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.462159 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-http-certs-internal\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.462355 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-remote-ca\"" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571818 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571863 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571890 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571910 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571928 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571947 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571975 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.571999 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.572027 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.572041 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.572063 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.572086 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.572100 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.572127 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/a0d396fc-0dad-43d0-902d-83606a2a7641-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.572144 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.672940 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.672992 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673030 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673050 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673080 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673110 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673133 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673166 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/a0d396fc-0dad-43d0-902d-83606a2a7641-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673195 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673241 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673268 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673301 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673327 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673355 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673883 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.673943 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.674146 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.674854 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.675840 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.675841 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.677087 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.679944 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.680388 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.680612 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/a0d396fc-0dad-43d0-902d-83606a2a7641-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.680617 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.680726 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.681239 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.682327 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.682651 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.689658 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/a0d396fc-0dad-43d0-902d-83606a2a7641-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"a0d396fc-0dad-43d0-902d-83606a2a7641\") " pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:50 crc kubenswrapper[5109]: I1211 17:06:50.775219 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:06:54 crc kubenswrapper[5109]: E1211 17:06:54.744043 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4 is running failed: container process not found" containerID="33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 17:06:54 crc kubenswrapper[5109]: E1211 17:06:54.744681 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4 is running failed: container process not found" containerID="33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 17:06:54 crc kubenswrapper[5109]: E1211 17:06:54.744955 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4 is running failed: container process not found" containerID="33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 17:06:54 crc kubenswrapper[5109]: E1211 17:06:54.744976 5109 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-mpfqq" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="registry-server" probeResult="unknown" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.218472 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.343160 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-utilities\") pod \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.343216 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dch5m\" (UniqueName: \"kubernetes.io/projected/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-kube-api-access-dch5m\") pod \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.343300 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-catalog-content\") pod \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\" (UID: \"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291\") " Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.353542 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-utilities" (OuterVolumeSpecName: "utilities") pod "1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" (UID: "1a9f4fb2-4653-4fcb-a2b8-57a3106b9291"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.379014 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-kube-api-access-dch5m" (OuterVolumeSpecName: "kube-api-access-dch5m") pod "1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" (UID: "1a9f4fb2-4653-4fcb-a2b8-57a3106b9291"). InnerVolumeSpecName "kube-api-access-dch5m". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.411538 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" (UID: "1a9f4fb2-4653-4fcb-a2b8-57a3106b9291"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.445889 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.445920 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.445929 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-dch5m\" (UniqueName: \"kubernetes.io/projected/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291-kube-api-access-dch5m\") on node \"crc\" DevicePath \"\"" Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.487041 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz"] Dec 11 17:06:55 crc kubenswrapper[5109]: W1211 17:06:55.509795 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc0e184e6_698a_40cc_ba2d_62e5e16b2980.slice/crio-0e6a99e03dc67e197924648d3b5455c957ced88741a16d142d2c2c6d7516233c WatchSource:0}: Error finding container 0e6a99e03dc67e197924648d3b5455c957ced88741a16d142d2c2c6d7516233c: Status 404 returned error can't find the container with id 0e6a99e03dc67e197924648d3b5455c957ced88741a16d142d2c2c6d7516233c Dec 11 17:06:55 crc kubenswrapper[5109]: I1211 17:06:55.633643 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 11 17:06:55 crc kubenswrapper[5109]: W1211 17:06:55.641402 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0d396fc_0dad_43d0_902d_83606a2a7641.slice/crio-ac29ec151ebebac25c5a6cf98b90fa2b15d990c58b251826e404478b7c65f5f5 WatchSource:0}: Error finding container ac29ec151ebebac25c5a6cf98b90fa2b15d990c58b251826e404478b7c65f5f5: Status 404 returned error can't find the container with id ac29ec151ebebac25c5a6cf98b90fa2b15d990c58b251826e404478b7c65f5f5 Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.034933 5109 generic.go:358] "Generic (PLEG): container finished" podID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerID="7c58f0643136a33ee6372dfe16b28f08d1545496f472b6fa61655dd08aa887f5" exitCode=0 Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.035012 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xc448" event={"ID":"ee0bc95f-e0a3-4ef8-aa90-71475f70929d","Type":"ContainerDied","Data":"7c58f0643136a33ee6372dfe16b28f08d1545496f472b6fa61655dd08aa887f5"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.037207 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" event={"ID":"116f2f7e-6d09-43af-9754-962ae5909372","Type":"ContainerStarted","Data":"c2106b0640f164d27d68c4192d8208b443a40f11036080f9b543843dd287e3c9"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.039387 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" event={"ID":"dd35d77f-5c0e-4a54-9a68-31e729d364bf","Type":"ContainerStarted","Data":"0e7b5d7be41016c15384128a27763bae073d3d510d7fe9908672fbfa46e147d9"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.040471 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.040522 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" event={"ID":"c0e184e6-698a-40cc-ba2d-62e5e16b2980","Type":"ContainerStarted","Data":"0e6a99e03dc67e197924648d3b5455c957ced88741a16d142d2c2c6d7516233c"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.042187 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" event={"ID":"c5432d60-106c-4c32-b8f4-e4064b8a71c8","Type":"ContainerStarted","Data":"92f7f8cf384daf40f1177912ad6c9c2e5ba319870688cd4f49c320d943abf0ff"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.044471 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mpfqq" event={"ID":"1a9f4fb2-4653-4fcb-a2b8-57a3106b9291","Type":"ContainerDied","Data":"87618a61602e4e32a3487b70bcb344ef05f1f9c41a22c081d6206a86017a9c58"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.044533 5109 scope.go:117] "RemoveContainer" containerID="33fdf64589b9e8b69935332af68a2c4bafaa7ba48a49d2905a06e83d650842a4" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.044544 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mpfqq" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.048895 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a0d396fc-0dad-43d0-902d-83606a2a7641","Type":"ContainerStarted","Data":"ac29ec151ebebac25c5a6cf98b90fa2b15d990c58b251826e404478b7c65f5f5"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.050689 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" event={"ID":"0ccdb48b-5a70-4b46-baf4-ae5a9d0fed77","Type":"ContainerStarted","Data":"3555052b0fbf0862b28bbf1ecb4e50d7c2cb51789d106d100b667b6f23ae3c5b"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.054930 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-78c97476f4-8rw8j" event={"ID":"042ee66e-ed7c-44e5-aa2d-bcc5bea13707","Type":"ContainerStarted","Data":"31a374ec5a46eb674bbe014c365545670822370d3107c7752db6f6baa7a5c9f1"} Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.055383 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.061431 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-78c97476f4-8rw8j" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.075329 5109 scope.go:117] "RemoveContainer" containerID="cfb5594ed0f858b5c0ac95a5fc1e5dbc514f990e93ab5c72b97fe35255ebb685" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.096433 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-86648f486b-h4fwb" podStartSLOduration=2.718081325 podStartE2EDuration="22.096415358s" podCreationTimestamp="2025-12-11 17:06:34 +0000 UTC" firstStartedPulling="2025-12-11 17:06:35.880073247 +0000 UTC m=+833.559764713" lastFinishedPulling="2025-12-11 17:06:55.25840728 +0000 UTC m=+852.938098746" observedRunningTime="2025-12-11 17:06:56.089874009 +0000 UTC m=+853.769565485" watchObservedRunningTime="2025-12-11 17:06:56.096415358 +0000 UTC m=+853.776106824" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.122913 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-78c97476f4-8rw8j" podStartSLOduration=4.989899867 podStartE2EDuration="21.122871475s" podCreationTimestamp="2025-12-11 17:06:35 +0000 UTC" firstStartedPulling="2025-12-11 17:06:39.161053952 +0000 UTC m=+836.840745418" lastFinishedPulling="2025-12-11 17:06:55.29402556 +0000 UTC m=+852.973717026" observedRunningTime="2025-12-11 17:06:56.108263138 +0000 UTC m=+853.787954634" watchObservedRunningTime="2025-12-11 17:06:56.122871475 +0000 UTC m=+853.802562941" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.125730 5109 scope.go:117] "RemoveContainer" containerID="44e234f6b7a03eb22d2656aabe010b94846d783cb0bb0b707be71d814c6786a3" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.206960 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-wz9sc" podStartSLOduration=2.841942098 podStartE2EDuration="21.206936538s" podCreationTimestamp="2025-12-11 17:06:35 +0000 UTC" firstStartedPulling="2025-12-11 17:06:36.9281907 +0000 UTC m=+834.607882166" lastFinishedPulling="2025-12-11 17:06:55.29318514 +0000 UTC m=+852.972876606" observedRunningTime="2025-12-11 17:06:56.16240084 +0000 UTC m=+853.842092306" watchObservedRunningTime="2025-12-11 17:06:56.206936538 +0000 UTC m=+853.886628004" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.208774 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" podStartSLOduration=6.735269538 podStartE2EDuration="21.208764403s" podCreationTimestamp="2025-12-11 17:06:35 +0000 UTC" firstStartedPulling="2025-12-11 17:06:40.784996267 +0000 UTC m=+838.464687733" lastFinishedPulling="2025-12-11 17:06:55.258491122 +0000 UTC m=+852.938182598" observedRunningTime="2025-12-11 17:06:56.194674588 +0000 UTC m=+853.874366054" watchObservedRunningTime="2025-12-11 17:06:56.208764403 +0000 UTC m=+853.888455879" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.234226 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-79d5b75947-jvjqt" podStartSLOduration=2.977527049 podStartE2EDuration="21.234202744s" podCreationTimestamp="2025-12-11 17:06:35 +0000 UTC" firstStartedPulling="2025-12-11 17:06:36.942076458 +0000 UTC m=+834.621767924" lastFinishedPulling="2025-12-11 17:06:55.198752153 +0000 UTC m=+852.878443619" observedRunningTime="2025-12-11 17:06:56.216135873 +0000 UTC m=+853.895827339" watchObservedRunningTime="2025-12-11 17:06:56.234202744 +0000 UTC m=+853.913894221" Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.251798 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mpfqq"] Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.260226 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mpfqq"] Dec 11 17:06:56 crc kubenswrapper[5109]: I1211 17:06:56.907955 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" path="/var/lib/kubelet/pods/1a9f4fb2-4653-4fcb-a2b8-57a3106b9291/volumes" Dec 11 17:06:57 crc kubenswrapper[5109]: I1211 17:06:57.071991 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xc448" event={"ID":"ee0bc95f-e0a3-4ef8-aa90-71475f70929d","Type":"ContainerStarted","Data":"0bbe6ad7da19c657de93d4a3c2e4089b3ed5edf23455ad0b21fc954509cb8735"} Dec 11 17:06:57 crc kubenswrapper[5109]: I1211 17:06:57.098976 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xc448" podStartSLOduration=4.706888731 podStartE2EDuration="18.098960076s" podCreationTimestamp="2025-12-11 17:06:39 +0000 UTC" firstStartedPulling="2025-12-11 17:06:41.901788371 +0000 UTC m=+839.581479837" lastFinishedPulling="2025-12-11 17:06:55.293859716 +0000 UTC m=+852.973551182" observedRunningTime="2025-12-11 17:06:57.094655011 +0000 UTC m=+854.774346477" watchObservedRunningTime="2025-12-11 17:06:57.098960076 +0000 UTC m=+854.778651542" Dec 11 17:06:59 crc kubenswrapper[5109]: I1211 17:06:59.667356 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" podUID="eead6019-41c4-4ac6-9226-8ec4c456db13" containerName="registry" containerID="cri-o://ddfcef0be238b0f543e5629758bb06f24ae93d271be9a1f78fadaba269b1b9de" gracePeriod=30 Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.008674 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-2ljml"] Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.009484 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="registry-server" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.009500 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="registry-server" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.009514 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="extract-content" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.009521 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="extract-content" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.009537 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="extract-utilities" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.009543 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="extract-utilities" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.009682 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="1a9f4fb2-4653-4fcb-a2b8-57a3106b9291" containerName="registry-server" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.015432 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.016473 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-2ljml"] Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.020506 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"infrawatch-operators-dockercfg-p4jxw\"" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.122020 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mfh7s\" (UniqueName: \"kubernetes.io/projected/0d5fd454-fe82-46d8-9fb9-a18ffe55043e-kube-api-access-mfh7s\") pod \"infrawatch-operators-2ljml\" (UID: \"0d5fd454-fe82-46d8-9fb9-a18ffe55043e\") " pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.128532 5109 generic.go:358] "Generic (PLEG): container finished" podID="eead6019-41c4-4ac6-9226-8ec4c456db13" containerID="ddfcef0be238b0f543e5629758bb06f24ae93d271be9a1f78fadaba269b1b9de" exitCode=0 Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.128634 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" event={"ID":"eead6019-41c4-4ac6-9226-8ec4c456db13","Type":"ContainerDied","Data":"ddfcef0be238b0f543e5629758bb06f24ae93d271be9a1f78fadaba269b1b9de"} Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.223392 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mfh7s\" (UniqueName: \"kubernetes.io/projected/0d5fd454-fe82-46d8-9fb9-a18ffe55043e-kube-api-access-mfh7s\") pod \"infrawatch-operators-2ljml\" (UID: \"0d5fd454-fe82-46d8-9fb9-a18ffe55043e\") " pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.253228 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mfh7s\" (UniqueName: \"kubernetes.io/projected/0d5fd454-fe82-46d8-9fb9-a18ffe55043e-kube-api-access-mfh7s\") pod \"infrawatch-operators-2ljml\" (UID: \"0d5fd454-fe82-46d8-9fb9-a18ffe55043e\") " pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.335289 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.789986 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.790036 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:07:00 crc kubenswrapper[5109]: I1211 17:07:00.862501 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:07:01 crc kubenswrapper[5109]: I1211 17:07:01.176434 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:07:02 crc kubenswrapper[5109]: I1211 17:07:02.594655 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xc448"] Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.070137 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.158687 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eead6019-41c4-4ac6-9226-8ec4c456db13-ca-trust-extracted\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.158980 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-trusted-ca\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159016 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8k4t\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-kube-api-access-r8k4t\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159487 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159557 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-certificates\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159593 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-bound-sa-token\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159625 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-tls\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159827 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" event={"ID":"eead6019-41c4-4ac6-9226-8ec4c456db13","Type":"ContainerDied","Data":"63ce238e0168bb3f1a6aa70488fb3da556a43d96b9d0bf1b56faddf8c18b07eb"} Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159895 5109 scope.go:117] "RemoveContainer" containerID="ddfcef0be238b0f543e5629758bb06f24ae93d271be9a1f78fadaba269b1b9de" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159940 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66587d64c8-nkxtk" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.159979 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xc448" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="registry-server" containerID="cri-o://0bbe6ad7da19c657de93d4a3c2e4089b3ed5edf23455ad0b21fc954509cb8735" gracePeriod=2 Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.160147 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.160178 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.160225 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eead6019-41c4-4ac6-9226-8ec4c456db13-installation-pull-secrets\") pod \"eead6019-41c4-4ac6-9226-8ec4c456db13\" (UID: \"eead6019-41c4-4ac6-9226-8ec4c456db13\") " Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.160490 5109 reconciler_common.go:299] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.160506 5109 reconciler_common.go:299] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/eead6019-41c4-4ac6-9226-8ec4c456db13-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.169761 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eead6019-41c4-4ac6-9226-8ec4c456db13-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.170374 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.170829 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-kube-api-access-r8k4t" (OuterVolumeSpecName: "kube-api-access-r8k4t") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "kube-api-access-r8k4t". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.178422 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.179368 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2" (OuterVolumeSpecName: "registry-storage") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "pvc-b21f41aa-58d4-44b1-aeaa-280a8e32ddf2". PluginName "kubernetes.io/csi", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.187454 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eead6019-41c4-4ac6-9226-8ec4c456db13-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "eead6019-41c4-4ac6-9226-8ec4c456db13" (UID: "eead6019-41c4-4ac6-9226-8ec4c456db13"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.262032 5109 reconciler_common.go:299] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.262232 5109 reconciler_common.go:299] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.262241 5109 reconciler_common.go:299] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/eead6019-41c4-4ac6-9226-8ec4c456db13-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.262251 5109 reconciler_common.go:299] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/eead6019-41c4-4ac6-9226-8ec4c456db13-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.262259 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-r8k4t\" (UniqueName: \"kubernetes.io/projected/eead6019-41c4-4ac6-9226-8ec4c456db13-kube-api-access-r8k4t\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.417432 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-2ljml"] Dec 11 17:07:03 crc kubenswrapper[5109]: W1211 17:07:03.429419 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d5fd454_fe82_46d8_9fb9_a18ffe55043e.slice/crio-f3599524e03060b287e43a681a5b8b6bfa6e42c7db658a0039d453660c374ed8 WatchSource:0}: Error finding container f3599524e03060b287e43a681a5b8b6bfa6e42c7db658a0039d453660c374ed8: Status 404 returned error can't find the container with id f3599524e03060b287e43a681a5b8b6bfa6e42c7db658a0039d453660c374ed8 Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.489475 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-nkxtk"] Dec 11 17:07:03 crc kubenswrapper[5109]: I1211 17:07:03.496406 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-66587d64c8-nkxtk"] Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.170944 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-2ljml" event={"ID":"0d5fd454-fe82-46d8-9fb9-a18ffe55043e","Type":"ContainerStarted","Data":"f3599524e03060b287e43a681a5b8b6bfa6e42c7db658a0039d453660c374ed8"} Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.172815 5109 generic.go:358] "Generic (PLEG): container finished" podID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerID="0bbe6ad7da19c657de93d4a3c2e4089b3ed5edf23455ad0b21fc954509cb8735" exitCode=0 Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.172878 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xc448" event={"ID":"ee0bc95f-e0a3-4ef8-aa90-71475f70929d","Type":"ContainerDied","Data":"0bbe6ad7da19c657de93d4a3c2e4089b3ed5edf23455ad0b21fc954509cb8735"} Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.176099 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" event={"ID":"c0e184e6-698a-40cc-ba2d-62e5e16b2980","Type":"ContainerStarted","Data":"7fa52663e9229225e83f0ff8d246cc9b840e8f3e956fdc4ff5e1f1762eeeb9b0"} Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.194957 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-64c74584c4-hdjzz" podStartSLOduration=10.688221373 podStartE2EDuration="18.194939859s" podCreationTimestamp="2025-12-11 17:06:46 +0000 UTC" firstStartedPulling="2025-12-11 17:06:55.526412946 +0000 UTC m=+853.206104422" lastFinishedPulling="2025-12-11 17:07:03.033131442 +0000 UTC m=+860.712822908" observedRunningTime="2025-12-11 17:07:04.193123224 +0000 UTC m=+861.872814700" watchObservedRunningTime="2025-12-11 17:07:04.194939859 +0000 UTC m=+861.874631325" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.781668 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.880128 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-utilities\") pod \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.880319 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-catalog-content\") pod \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.880351 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2bws\" (UniqueName: \"kubernetes.io/projected/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-kube-api-access-n2bws\") pod \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\" (UID: \"ee0bc95f-e0a3-4ef8-aa90-71475f70929d\") " Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.882792 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-utilities" (OuterVolumeSpecName: "utilities") pod "ee0bc95f-e0a3-4ef8-aa90-71475f70929d" (UID: "ee0bc95f-e0a3-4ef8-aa90-71475f70929d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.891922 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-kube-api-access-n2bws" (OuterVolumeSpecName: "kube-api-access-n2bws") pod "ee0bc95f-e0a3-4ef8-aa90-71475f70929d" (UID: "ee0bc95f-e0a3-4ef8-aa90-71475f70929d"). InnerVolumeSpecName "kube-api-access-n2bws". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.923012 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eead6019-41c4-4ac6-9226-8ec4c456db13" path="/var/lib/kubelet/pods/eead6019-41c4-4ac6-9226-8ec4c456db13/volumes" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.955464 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee0bc95f-e0a3-4ef8-aa90-71475f70929d" (UID: "ee0bc95f-e0a3-4ef8-aa90-71475f70929d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.982119 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.982154 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:04 crc kubenswrapper[5109]: I1211 17:07:04.982166 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-n2bws\" (UniqueName: \"kubernetes.io/projected/ee0bc95f-e0a3-4ef8-aa90-71475f70929d-kube-api-access-n2bws\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:05 crc kubenswrapper[5109]: I1211 17:07:05.185530 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xc448" event={"ID":"ee0bc95f-e0a3-4ef8-aa90-71475f70929d","Type":"ContainerDied","Data":"3efa8c5c650625d5cc8fd088a8d3c0d5f388de99b8b29d6849c0be33936a43e3"} Dec 11 17:07:05 crc kubenswrapper[5109]: I1211 17:07:05.185578 5109 scope.go:117] "RemoveContainer" containerID="0bbe6ad7da19c657de93d4a3c2e4089b3ed5edf23455ad0b21fc954509cb8735" Dec 11 17:07:05 crc kubenswrapper[5109]: I1211 17:07:05.185764 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xc448" Dec 11 17:07:05 crc kubenswrapper[5109]: I1211 17:07:05.209812 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xc448"] Dec 11 17:07:05 crc kubenswrapper[5109]: I1211 17:07:05.213059 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xc448"] Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.647640 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-7894b5b9b4-q64b2"] Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648534 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="eead6019-41c4-4ac6-9226-8ec4c456db13" containerName="registry" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648549 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="eead6019-41c4-4ac6-9226-8ec4c456db13" containerName="registry" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648571 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="extract-content" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648579 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="extract-content" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648606 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="registry-server" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648614 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="registry-server" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648631 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="extract-utilities" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648638 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="extract-utilities" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648775 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" containerName="registry-server" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.648796 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="eead6019-41c4-4ac6-9226-8ec4c456db13" containerName="registry" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.695817 5109 scope.go:117] "RemoveContainer" containerID="7c58f0643136a33ee6372dfe16b28f08d1545496f472b6fa61655dd08aa887f5" Dec 11 17:07:06 crc kubenswrapper[5109]: I1211 17:07:06.712512 5109 scope.go:117] "RemoveContainer" containerID="522c2492c160a787504995fa3f8058b169438653e309eb226d61b9ec164f9a0a" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.161972 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.165574 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"openshift-service-ca.crt\"" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.165839 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"cert-manager\"/\"kube-root-ca.crt\"" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.167058 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-webhook-dockercfg-58kc4\"" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.172827 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee0bc95f-e0a3-4ef8-aa90-71475f70929d" path="/var/lib/kubelet/pods/ee0bc95f-e0a3-4ef8-aa90-71475f70929d/volumes" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.173491 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-7894b5b9b4-q64b2"] Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.173580 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-68bdb49cbf-sqjbr" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.210349 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2bxp\" (UniqueName: \"kubernetes.io/projected/87f20cd3-e95a-499a-859e-a4c7a1d984de-kube-api-access-j2bxp\") pod \"cert-manager-webhook-7894b5b9b4-q64b2\" (UID: \"87f20cd3-e95a-499a-859e-a4c7a1d984de\") " pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.210430 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/87f20cd3-e95a-499a-859e-a4c7a1d984de-bound-sa-token\") pod \"cert-manager-webhook-7894b5b9b4-q64b2\" (UID: \"87f20cd3-e95a-499a-859e-a4c7a1d984de\") " pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.311422 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-j2bxp\" (UniqueName: \"kubernetes.io/projected/87f20cd3-e95a-499a-859e-a4c7a1d984de-kube-api-access-j2bxp\") pod \"cert-manager-webhook-7894b5b9b4-q64b2\" (UID: \"87f20cd3-e95a-499a-859e-a4c7a1d984de\") " pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.311682 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/87f20cd3-e95a-499a-859e-a4c7a1d984de-bound-sa-token\") pod \"cert-manager-webhook-7894b5b9b4-q64b2\" (UID: \"87f20cd3-e95a-499a-859e-a4c7a1d984de\") " pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.329654 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/87f20cd3-e95a-499a-859e-a4c7a1d984de-bound-sa-token\") pod \"cert-manager-webhook-7894b5b9b4-q64b2\" (UID: \"87f20cd3-e95a-499a-859e-a4c7a1d984de\") " pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.348048 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2bxp\" (UniqueName: \"kubernetes.io/projected/87f20cd3-e95a-499a-859e-a4c7a1d984de-kube-api-access-j2bxp\") pod \"cert-manager-webhook-7894b5b9b4-q64b2\" (UID: \"87f20cd3-e95a-499a-859e-a4c7a1d984de\") " pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.479533 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.761769 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:07:07 crc kubenswrapper[5109]: I1211 17:07:07.761862 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:07:10 crc kubenswrapper[5109]: I1211 17:07:10.176436 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6"] Dec 11 17:07:10 crc kubenswrapper[5109]: I1211 17:07:10.928712 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:10 crc kubenswrapper[5109]: I1211 17:07:10.930205 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-cainjector-dockercfg-dprl6\"" Dec 11 17:07:10 crc kubenswrapper[5109]: I1211 17:07:10.935397 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6"] Dec 11 17:07:10 crc kubenswrapper[5109]: I1211 17:07:10.963140 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rg8ww\" (UniqueName: \"kubernetes.io/projected/b23da9d4-1b3c-472a-af3f-35111370df37-kube-api-access-rg8ww\") pod \"cert-manager-cainjector-7dbf76d5c8-rhjz6\" (UID: \"b23da9d4-1b3c-472a-af3f-35111370df37\") " pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:10 crc kubenswrapper[5109]: I1211 17:07:10.963251 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b23da9d4-1b3c-472a-af3f-35111370df37-bound-sa-token\") pod \"cert-manager-cainjector-7dbf76d5c8-rhjz6\" (UID: \"b23da9d4-1b3c-472a-af3f-35111370df37\") " pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:11 crc kubenswrapper[5109]: I1211 17:07:11.064613 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b23da9d4-1b3c-472a-af3f-35111370df37-bound-sa-token\") pod \"cert-manager-cainjector-7dbf76d5c8-rhjz6\" (UID: \"b23da9d4-1b3c-472a-af3f-35111370df37\") " pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:11 crc kubenswrapper[5109]: I1211 17:07:11.064754 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-rg8ww\" (UniqueName: \"kubernetes.io/projected/b23da9d4-1b3c-472a-af3f-35111370df37-kube-api-access-rg8ww\") pod \"cert-manager-cainjector-7dbf76d5c8-rhjz6\" (UID: \"b23da9d4-1b3c-472a-af3f-35111370df37\") " pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:11 crc kubenswrapper[5109]: I1211 17:07:11.091936 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-rg8ww\" (UniqueName: \"kubernetes.io/projected/b23da9d4-1b3c-472a-af3f-35111370df37-kube-api-access-rg8ww\") pod \"cert-manager-cainjector-7dbf76d5c8-rhjz6\" (UID: \"b23da9d4-1b3c-472a-af3f-35111370df37\") " pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:11 crc kubenswrapper[5109]: I1211 17:07:11.103194 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b23da9d4-1b3c-472a-af3f-35111370df37-bound-sa-token\") pod \"cert-manager-cainjector-7dbf76d5c8-rhjz6\" (UID: \"b23da9d4-1b3c-472a-af3f-35111370df37\") " pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:11 crc kubenswrapper[5109]: I1211 17:07:11.248550 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" Dec 11 17:07:18 crc kubenswrapper[5109]: I1211 17:07:18.830184 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858d87f86b-jsnrq"] Dec 11 17:07:19 crc kubenswrapper[5109]: I1211 17:07:19.993165 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.000821 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"cert-manager\"/\"cert-manager-dockercfg-rf2pc\"" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.007021 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858d87f86b-jsnrq"] Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.147654 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5250110d-c32c-4715-8605-1a181fad0918-bound-sa-token\") pod \"cert-manager-858d87f86b-jsnrq\" (UID: \"5250110d-c32c-4715-8605-1a181fad0918\") " pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.149236 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbqsq\" (UniqueName: \"kubernetes.io/projected/5250110d-c32c-4715-8605-1a181fad0918-kube-api-access-wbqsq\") pod \"cert-manager-858d87f86b-jsnrq\" (UID: \"5250110d-c32c-4715-8605-1a181fad0918\") " pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.250620 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wbqsq\" (UniqueName: \"kubernetes.io/projected/5250110d-c32c-4715-8605-1a181fad0918-kube-api-access-wbqsq\") pod \"cert-manager-858d87f86b-jsnrq\" (UID: \"5250110d-c32c-4715-8605-1a181fad0918\") " pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.250697 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5250110d-c32c-4715-8605-1a181fad0918-bound-sa-token\") pod \"cert-manager-858d87f86b-jsnrq\" (UID: \"5250110d-c32c-4715-8605-1a181fad0918\") " pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.270792 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbqsq\" (UniqueName: \"kubernetes.io/projected/5250110d-c32c-4715-8605-1a181fad0918-kube-api-access-wbqsq\") pod \"cert-manager-858d87f86b-jsnrq\" (UID: \"5250110d-c32c-4715-8605-1a181fad0918\") " pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.271714 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/5250110d-c32c-4715-8605-1a181fad0918-bound-sa-token\") pod \"cert-manager-858d87f86b-jsnrq\" (UID: \"5250110d-c32c-4715-8605-1a181fad0918\") " pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.319620 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858d87f86b-jsnrq" Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.695257 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-7894b5b9b4-q64b2"] Dec 11 17:07:20 crc kubenswrapper[5109]: W1211 17:07:20.754644 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87f20cd3_e95a_499a_859e_a4c7a1d984de.slice/crio-d569bfd474885c45f2e2aa53d58a37ad0631bb8085c92febc6b0ecea73c241b7 WatchSource:0}: Error finding container d569bfd474885c45f2e2aa53d58a37ad0631bb8085c92febc6b0ecea73c241b7: Status 404 returned error can't find the container with id d569bfd474885c45f2e2aa53d58a37ad0631bb8085c92febc6b0ecea73c241b7 Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.963091 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6"] Dec 11 17:07:20 crc kubenswrapper[5109]: W1211 17:07:20.975179 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb23da9d4_1b3c_472a_af3f_35111370df37.slice/crio-aac5636e59c097c03b108e5add20e5f1095a457a0a26ad1a10f4bfef145fe39c WatchSource:0}: Error finding container aac5636e59c097c03b108e5add20e5f1095a457a0a26ad1a10f4bfef145fe39c: Status 404 returned error can't find the container with id aac5636e59c097c03b108e5add20e5f1095a457a0a26ad1a10f4bfef145fe39c Dec 11 17:07:20 crc kubenswrapper[5109]: I1211 17:07:20.992377 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858d87f86b-jsnrq"] Dec 11 17:07:20 crc kubenswrapper[5109]: W1211 17:07:20.995110 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5250110d_c32c_4715_8605_1a181fad0918.slice/crio-c1e0dfd6e35cccccf9d45c349ec126d61ffa2af07dc246f0228478bd0a9f8ca2 WatchSource:0}: Error finding container c1e0dfd6e35cccccf9d45c349ec126d61ffa2af07dc246f0228478bd0a9f8ca2: Status 404 returned error can't find the container with id c1e0dfd6e35cccccf9d45c349ec126d61ffa2af07dc246f0228478bd0a9f8ca2 Dec 11 17:07:21 crc kubenswrapper[5109]: I1211 17:07:21.330971 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a0d396fc-0dad-43d0-902d-83606a2a7641","Type":"ContainerStarted","Data":"d11ee5f824b01624cc4c73bac5d3773c740cea9f6598a72709bb97cf62856628"} Dec 11 17:07:21 crc kubenswrapper[5109]: I1211 17:07:21.332352 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" event={"ID":"b23da9d4-1b3c-472a-af3f-35111370df37","Type":"ContainerStarted","Data":"aac5636e59c097c03b108e5add20e5f1095a457a0a26ad1a10f4bfef145fe39c"} Dec 11 17:07:21 crc kubenswrapper[5109]: I1211 17:07:21.333633 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" event={"ID":"87f20cd3-e95a-499a-859e-a4c7a1d984de","Type":"ContainerStarted","Data":"d569bfd474885c45f2e2aa53d58a37ad0631bb8085c92febc6b0ecea73c241b7"} Dec 11 17:07:21 crc kubenswrapper[5109]: I1211 17:07:21.335069 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858d87f86b-jsnrq" event={"ID":"5250110d-c32c-4715-8605-1a181fad0918","Type":"ContainerStarted","Data":"c1e0dfd6e35cccccf9d45c349ec126d61ffa2af07dc246f0228478bd0a9f8ca2"} Dec 11 17:07:21 crc kubenswrapper[5109]: I1211 17:07:21.556515 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 11 17:07:21 crc kubenswrapper[5109]: I1211 17:07:21.594396 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Dec 11 17:07:25 crc kubenswrapper[5109]: I1211 17:07:25.367261 5109 generic.go:358] "Generic (PLEG): container finished" podID="a0d396fc-0dad-43d0-902d-83606a2a7641" containerID="d11ee5f824b01624cc4c73bac5d3773c740cea9f6598a72709bb97cf62856628" exitCode=0 Dec 11 17:07:25 crc kubenswrapper[5109]: I1211 17:07:25.367451 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a0d396fc-0dad-43d0-902d-83606a2a7641","Type":"ContainerDied","Data":"d11ee5f824b01624cc4c73bac5d3773c740cea9f6598a72709bb97cf62856628"} Dec 11 17:07:30 crc kubenswrapper[5109]: I1211 17:07:30.409705 5109 generic.go:358] "Generic (PLEG): container finished" podID="a0d396fc-0dad-43d0-902d-83606a2a7641" containerID="66fee956dda36f4ddb431a1a94d6e1509e6db8dacd3f87c1ff7a34c0f462ab36" exitCode=0 Dec 11 17:07:30 crc kubenswrapper[5109]: I1211 17:07:30.409773 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a0d396fc-0dad-43d0-902d-83606a2a7641","Type":"ContainerDied","Data":"66fee956dda36f4ddb431a1a94d6e1509e6db8dacd3f87c1ff7a34c0f462ab36"} Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.421530 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"a0d396fc-0dad-43d0-902d-83606a2a7641","Type":"ContainerStarted","Data":"a2d3791133e94f8737e6e2c9baef05b9e58c4717679d60e1ec3fd35e4b5b96eb"} Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.423908 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.426356 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" event={"ID":"b23da9d4-1b3c-472a-af3f-35111370df37","Type":"ContainerStarted","Data":"84f05dd2063d74820f82f8bcbf10d0db2d84b260fa6ea28ddc73b880fd1344ba"} Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.455206 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" event={"ID":"87f20cd3-e95a-499a-859e-a4c7a1d984de","Type":"ContainerStarted","Data":"b359665526d6a6c069d3e2279f8c83231641fdbc687990a296e8420eea9574f4"} Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.455659 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.457586 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858d87f86b-jsnrq" event={"ID":"5250110d-c32c-4715-8605-1a181fad0918","Type":"ContainerStarted","Data":"c982ae847b6ac47b5f4de9490d4928cb8b3a1f85547a4eadcc97e3372f4ee158"} Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.464359 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-2ljml" event={"ID":"0d5fd454-fe82-46d8-9fb9-a18ffe55043e","Type":"ContainerStarted","Data":"b4602eb90d0d3a3e32ac4cc4ce0696158c75f3caf7bbe5a093c267adf3539290"} Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.486318 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=16.097183421 podStartE2EDuration="41.486301844s" podCreationTimestamp="2025-12-11 17:06:50 +0000 UTC" firstStartedPulling="2025-12-11 17:06:55.644216473 +0000 UTC m=+853.323907939" lastFinishedPulling="2025-12-11 17:07:21.033334896 +0000 UTC m=+878.713026362" observedRunningTime="2025-12-11 17:07:31.483785852 +0000 UTC m=+889.163477328" watchObservedRunningTime="2025-12-11 17:07:31.486301844 +0000 UTC m=+889.165993320" Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.505916 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-2ljml" podStartSLOduration=5.900792479 podStartE2EDuration="32.505895072s" podCreationTimestamp="2025-12-11 17:06:59 +0000 UTC" firstStartedPulling="2025-12-11 17:07:03.430613399 +0000 UTC m=+861.110304865" lastFinishedPulling="2025-12-11 17:07:30.035715992 +0000 UTC m=+887.715407458" observedRunningTime="2025-12-11 17:07:31.504792135 +0000 UTC m=+889.184483631" watchObservedRunningTime="2025-12-11 17:07:31.505895072 +0000 UTC m=+889.185586578" Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.535881 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858d87f86b-jsnrq" podStartSLOduration=4.232375443 podStartE2EDuration="13.535863225s" podCreationTimestamp="2025-12-11 17:07:18 +0000 UTC" firstStartedPulling="2025-12-11 17:07:20.997328826 +0000 UTC m=+878.677020292" lastFinishedPulling="2025-12-11 17:07:30.300816608 +0000 UTC m=+887.980508074" observedRunningTime="2025-12-11 17:07:31.531098988 +0000 UTC m=+889.210790464" watchObservedRunningTime="2025-12-11 17:07:31.535863225 +0000 UTC m=+889.215554691" Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.566881 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7dbf76d5c8-rhjz6" podStartSLOduration=12.010856602 podStartE2EDuration="21.566864002s" podCreationTimestamp="2025-12-11 17:07:10 +0000 UTC" firstStartedPulling="2025-12-11 17:07:20.97908577 +0000 UTC m=+878.658777236" lastFinishedPulling="2025-12-11 17:07:30.53509316 +0000 UTC m=+888.214784636" observedRunningTime="2025-12-11 17:07:31.562773672 +0000 UTC m=+889.242465138" watchObservedRunningTime="2025-12-11 17:07:31.566864002 +0000 UTC m=+889.246555468" Dec 11 17:07:31 crc kubenswrapper[5109]: I1211 17:07:31.587642 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" podStartSLOduration=15.972461855 podStartE2EDuration="25.587625919s" podCreationTimestamp="2025-12-11 17:07:06 +0000 UTC" firstStartedPulling="2025-12-11 17:07:20.757452207 +0000 UTC m=+878.437143673" lastFinishedPulling="2025-12-11 17:07:30.372616271 +0000 UTC m=+888.052307737" observedRunningTime="2025-12-11 17:07:31.581952891 +0000 UTC m=+889.261644357" watchObservedRunningTime="2025-12-11 17:07:31.587625919 +0000 UTC m=+889.267317385" Dec 11 17:07:37 crc kubenswrapper[5109]: I1211 17:07:37.474178 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-7894b5b9b4-q64b2" Dec 11 17:07:37 crc kubenswrapper[5109]: I1211 17:07:37.761750 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:07:37 crc kubenswrapper[5109]: I1211 17:07:37.761826 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:07:40 crc kubenswrapper[5109]: I1211 17:07:40.336287 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:40 crc kubenswrapper[5109]: I1211 17:07:40.336341 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:40 crc kubenswrapper[5109]: I1211 17:07:40.370466 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:40 crc kubenswrapper[5109]: I1211 17:07:40.558502 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-2ljml" Dec 11 17:07:42 crc kubenswrapper[5109]: I1211 17:07:42.546235 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="a0d396fc-0dad-43d0-902d-83606a2a7641" containerName="elasticsearch" probeResult="failure" output=< Dec 11 17:07:42 crc kubenswrapper[5109]: {"timestamp": "2025-12-11T17:07:42+00:00", "message": "readiness probe failed", "curl_rc": "7"} Dec 11 17:07:42 crc kubenswrapper[5109]: > Dec 11 17:07:43 crc kubenswrapper[5109]: I1211 17:07:43.248278 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 17:07:43 crc kubenswrapper[5109]: I1211 17:07:43.250178 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 17:07:43 crc kubenswrapper[5109]: I1211 17:07:43.270246 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hx2kb_7c45b4ae-02aa-4614-977b-544a9e303bdf/kube-multus/0.log" Dec 11 17:07:43 crc kubenswrapper[5109]: I1211 17:07:43.272960 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hx2kb_7c45b4ae-02aa-4614-977b-544a9e303bdf/kube-multus/0.log" Dec 11 17:07:43 crc kubenswrapper[5109]: I1211 17:07:43.278470 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 17:07:43 crc kubenswrapper[5109]: I1211 17:07:43.280711 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 17:07:47 crc kubenswrapper[5109]: I1211 17:07:47.526518 5109 prober.go:120] "Probe failed" probeType="Readiness" pod="service-telemetry/elasticsearch-es-default-0" podUID="a0d396fc-0dad-43d0-902d-83606a2a7641" containerName="elasticsearch" probeResult="failure" output=< Dec 11 17:07:47 crc kubenswrapper[5109]: {"timestamp": "2025-12-11T17:07:47+00:00", "message": "readiness probe failed", "curl_rc": "7"} Dec 11 17:07:47 crc kubenswrapper[5109]: > Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.497300 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4"] Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.502932 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.510579 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4"] Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.541788 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-bundle\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.541847 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-util\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.541928 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzk4r\" (UniqueName: \"kubernetes.io/projected/a88b7761-29b7-4a4c-ab01-951bfec5cad1-kube-api-access-zzk4r\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.643646 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zzk4r\" (UniqueName: \"kubernetes.io/projected/a88b7761-29b7-4a4c-ab01-951bfec5cad1-kube-api-access-zzk4r\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.643755 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-bundle\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.643777 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-util\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.644230 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-util\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.644253 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-bundle\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.664799 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzk4r\" (UniqueName: \"kubernetes.io/projected/a88b7761-29b7-4a4c-ab01-951bfec5cad1-kube-api-access-zzk4r\") pod \"357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:50 crc kubenswrapper[5109]: I1211 17:07:50.818951 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.057370 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4"] Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.298416 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn"] Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.302692 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.311414 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn"] Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.362361 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fb6b\" (UniqueName: \"kubernetes.io/projected/3890f761-4b7f-4033-8397-3b3639ee86a7-kube-api-access-9fb6b\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.362434 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-util\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.362482 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-bundle\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.463500 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-util\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.463641 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-bundle\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.463802 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-9fb6b\" (UniqueName: \"kubernetes.io/projected/3890f761-4b7f-4033-8397-3b3639ee86a7-kube-api-access-9fb6b\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.464019 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-util\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.464187 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-bundle\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.483436 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fb6b\" (UniqueName: \"kubernetes.io/projected/3890f761-4b7f-4033-8397-3b3639ee86a7-kube-api-access-9fb6b\") pod \"ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.634102 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.717712 5109 generic.go:358] "Generic (PLEG): container finished" podID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerID="d8278fab72cc8ed6f8ffaaa8ff90a348d9d7a362baec595a4328624caf474abf" exitCode=0 Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.717903 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" event={"ID":"a88b7761-29b7-4a4c-ab01-951bfec5cad1","Type":"ContainerDied","Data":"d8278fab72cc8ed6f8ffaaa8ff90a348d9d7a362baec595a4328624caf474abf"} Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.718030 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" event={"ID":"a88b7761-29b7-4a4c-ab01-951bfec5cad1","Type":"ContainerStarted","Data":"5e701f4796b2c8d6f9908076f765895ee0b154f5ef04de21ca49939eebfc73c8"} Dec 11 17:07:51 crc kubenswrapper[5109]: I1211 17:07:51.880988 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn"] Dec 11 17:07:51 crc kubenswrapper[5109]: W1211 17:07:51.881593 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3890f761_4b7f_4033_8397_3b3639ee86a7.slice/crio-dd5f95bfaa6d0c59af4fbbb2b1e3b04044862342fe1cdd55b6c611d158fa51a7 WatchSource:0}: Error finding container dd5f95bfaa6d0c59af4fbbb2b1e3b04044862342fe1cdd55b6c611d158fa51a7: Status 404 returned error can't find the container with id dd5f95bfaa6d0c59af4fbbb2b1e3b04044862342fe1cdd55b6c611d158fa51a7 Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.117571 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222"] Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.125235 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.129132 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222"] Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.131114 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-marketplace\"/\"default-dockercfg-b2ccr\"" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.172267 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.172364 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.172478 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzfcw\" (UniqueName: \"kubernetes.io/projected/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-kube-api-access-wzfcw\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.273459 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-wzfcw\" (UniqueName: \"kubernetes.io/projected/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-kube-api-access-wzfcw\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.273519 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.273618 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.274149 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.274235 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.302751 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzfcw\" (UniqueName: \"kubernetes.io/projected/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-kube-api-access-wzfcw\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.441540 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.717143 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.734593 5109 generic.go:358] "Generic (PLEG): container finished" podID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerID="ff4edc2e72680a7d81dc8bc5aa1431718857676694729e6d3420feac5f91152b" exitCode=0 Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.734769 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" event={"ID":"3890f761-4b7f-4033-8397-3b3639ee86a7","Type":"ContainerDied","Data":"ff4edc2e72680a7d81dc8bc5aa1431718857676694729e6d3420feac5f91152b"} Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.734804 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" event={"ID":"3890f761-4b7f-4033-8397-3b3639ee86a7","Type":"ContainerStarted","Data":"dd5f95bfaa6d0c59af4fbbb2b1e3b04044862342fe1cdd55b6c611d158fa51a7"} Dec 11 17:07:52 crc kubenswrapper[5109]: I1211 17:07:52.770764 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222"] Dec 11 17:07:53 crc kubenswrapper[5109]: I1211 17:07:53.742957 5109 generic.go:358] "Generic (PLEG): container finished" podID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerID="ff5372b520e5c262d60c4a8bc9556328e51779e116ae2d65064657dd034edfe8" exitCode=0 Dec 11 17:07:53 crc kubenswrapper[5109]: I1211 17:07:53.743305 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" event={"ID":"a88b7761-29b7-4a4c-ab01-951bfec5cad1","Type":"ContainerDied","Data":"ff5372b520e5c262d60c4a8bc9556328e51779e116ae2d65064657dd034edfe8"} Dec 11 17:07:53 crc kubenswrapper[5109]: I1211 17:07:53.745699 5109 generic.go:358] "Generic (PLEG): container finished" podID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerID="2641649b82c8d9c7678c01c24857dd092735aab86db032ffd7008a79e8cfb3ea" exitCode=0 Dec 11 17:07:53 crc kubenswrapper[5109]: I1211 17:07:53.745800 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" event={"ID":"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24","Type":"ContainerDied","Data":"2641649b82c8d9c7678c01c24857dd092735aab86db032ffd7008a79e8cfb3ea"} Dec 11 17:07:53 crc kubenswrapper[5109]: I1211 17:07:53.745829 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" event={"ID":"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24","Type":"ContainerStarted","Data":"52f729a705da6daa4427aa683099630e622ef45f70cded847d5088320593cd86"} Dec 11 17:07:54 crc kubenswrapper[5109]: I1211 17:07:54.756349 5109 generic.go:358] "Generic (PLEG): container finished" podID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerID="0bb77ad74c136bb110dac1c2f8a438d59de3d796def73c0d9b837be4a67ebd84" exitCode=0 Dec 11 17:07:54 crc kubenswrapper[5109]: I1211 17:07:54.756406 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" event={"ID":"a88b7761-29b7-4a4c-ab01-951bfec5cad1","Type":"ContainerDied","Data":"0bb77ad74c136bb110dac1c2f8a438d59de3d796def73c0d9b837be4a67ebd84"} Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.041625 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.132118 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzk4r\" (UniqueName: \"kubernetes.io/projected/a88b7761-29b7-4a4c-ab01-951bfec5cad1-kube-api-access-zzk4r\") pod \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.132240 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-bundle\") pod \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.132390 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-util\") pod \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\" (UID: \"a88b7761-29b7-4a4c-ab01-951bfec5cad1\") " Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.132705 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-bundle" (OuterVolumeSpecName: "bundle") pod "a88b7761-29b7-4a4c-ab01-951bfec5cad1" (UID: "a88b7761-29b7-4a4c-ab01-951bfec5cad1"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.132995 5109 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.144852 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88b7761-29b7-4a4c-ab01-951bfec5cad1-kube-api-access-zzk4r" (OuterVolumeSpecName: "kube-api-access-zzk4r") pod "a88b7761-29b7-4a4c-ab01-951bfec5cad1" (UID: "a88b7761-29b7-4a4c-ab01-951bfec5cad1"). InnerVolumeSpecName "kube-api-access-zzk4r". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.148005 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-util" (OuterVolumeSpecName: "util") pod "a88b7761-29b7-4a4c-ab01-951bfec5cad1" (UID: "a88b7761-29b7-4a4c-ab01-951bfec5cad1"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.235055 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-zzk4r\" (UniqueName: \"kubernetes.io/projected/a88b7761-29b7-4a4c-ab01-951bfec5cad1-kube-api-access-zzk4r\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.235096 5109 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a88b7761-29b7-4a4c-ab01-951bfec5cad1-util\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.771488 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.771493 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/357dd3a7f1018da1829ef8b7f45be6b50a1f8f645849b6e661ccd9d7684c4q4" event={"ID":"a88b7761-29b7-4a4c-ab01-951bfec5cad1","Type":"ContainerDied","Data":"5e701f4796b2c8d6f9908076f765895ee0b154f5ef04de21ca49939eebfc73c8"} Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.771558 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e701f4796b2c8d6f9908076f765895ee0b154f5ef04de21ca49939eebfc73c8" Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.774082 5109 generic.go:358] "Generic (PLEG): container finished" podID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerID="2140565c6cd17277b24ed2a872e8231c9600ff558bde6660c7f726cfcec421a1" exitCode=0 Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.774133 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" event={"ID":"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24","Type":"ContainerDied","Data":"2140565c6cd17277b24ed2a872e8231c9600ff558bde6660c7f726cfcec421a1"} Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.778044 5109 generic.go:358] "Generic (PLEG): container finished" podID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerID="f8c6eda7f3a2d0b60e90d646d0941403cffdeb0a503790fa26458813e4f0a007" exitCode=0 Dec 11 17:07:56 crc kubenswrapper[5109]: I1211 17:07:56.778104 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" event={"ID":"3890f761-4b7f-4033-8397-3b3639ee86a7","Type":"ContainerDied","Data":"f8c6eda7f3a2d0b60e90d646d0941403cffdeb0a503790fa26458813e4f0a007"} Dec 11 17:07:57 crc kubenswrapper[5109]: I1211 17:07:57.787595 5109 generic.go:358] "Generic (PLEG): container finished" podID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerID="42252938efed559f01f534179706f0813e4b515a05b4b3b0e0acfdfd38d6855b" exitCode=0 Dec 11 17:07:57 crc kubenswrapper[5109]: I1211 17:07:57.787676 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" event={"ID":"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24","Type":"ContainerDied","Data":"42252938efed559f01f534179706f0813e4b515a05b4b3b0e0acfdfd38d6855b"} Dec 11 17:07:57 crc kubenswrapper[5109]: I1211 17:07:57.791169 5109 generic.go:358] "Generic (PLEG): container finished" podID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerID="9d22620f7045072e608a0703210735a01929fddced2a102fc4b439531dd9307d" exitCode=0 Dec 11 17:07:57 crc kubenswrapper[5109]: I1211 17:07:57.791281 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" event={"ID":"3890f761-4b7f-4033-8397-3b3639ee86a7","Type":"ContainerDied","Data":"9d22620f7045072e608a0703210735a01929fddced2a102fc4b439531dd9307d"} Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.108137 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.118177 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.181505 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-bundle\") pod \"3890f761-4b7f-4033-8397-3b3639ee86a7\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.181561 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9fb6b\" (UniqueName: \"kubernetes.io/projected/3890f761-4b7f-4033-8397-3b3639ee86a7-kube-api-access-9fb6b\") pod \"3890f761-4b7f-4033-8397-3b3639ee86a7\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.181583 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-bundle\") pod \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.181644 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-util\") pod \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.181687 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzfcw\" (UniqueName: \"kubernetes.io/projected/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-kube-api-access-wzfcw\") pod \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\" (UID: \"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24\") " Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.181776 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-util\") pod \"3890f761-4b7f-4033-8397-3b3639ee86a7\" (UID: \"3890f761-4b7f-4033-8397-3b3639ee86a7\") " Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.182260 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-bundle" (OuterVolumeSpecName: "bundle") pod "3890f761-4b7f-4033-8397-3b3639ee86a7" (UID: "3890f761-4b7f-4033-8397-3b3639ee86a7"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.183568 5109 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.184477 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-bundle" (OuterVolumeSpecName: "bundle") pod "2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" (UID: "2d9dffb1-a917-4fae-b276-4e8bdbbf4d24"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.187510 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-kube-api-access-wzfcw" (OuterVolumeSpecName: "kube-api-access-wzfcw") pod "2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" (UID: "2d9dffb1-a917-4fae-b276-4e8bdbbf4d24"). InnerVolumeSpecName "kube-api-access-wzfcw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.188715 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3890f761-4b7f-4033-8397-3b3639ee86a7-kube-api-access-9fb6b" (OuterVolumeSpecName: "kube-api-access-9fb6b") pod "3890f761-4b7f-4033-8397-3b3639ee86a7" (UID: "3890f761-4b7f-4033-8397-3b3639ee86a7"). InnerVolumeSpecName "kube-api-access-9fb6b". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.192168 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-util" (OuterVolumeSpecName: "util") pod "2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" (UID: "2d9dffb1-a917-4fae-b276-4e8bdbbf4d24"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.193280 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-util" (OuterVolumeSpecName: "util") pod "3890f761-4b7f-4033-8397-3b3639ee86a7" (UID: "3890f761-4b7f-4033-8397-3b3639ee86a7"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.284454 5109 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3890f761-4b7f-4033-8397-3b3639ee86a7-util\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.284485 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-9fb6b\" (UniqueName: \"kubernetes.io/projected/3890f761-4b7f-4033-8397-3b3639ee86a7-kube-api-access-9fb6b\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.284495 5109 reconciler_common.go:299] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-bundle\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.284508 5109 reconciler_common.go:299] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-util\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.284516 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-wzfcw\" (UniqueName: \"kubernetes.io/projected/2d9dffb1-a917-4fae-b276-4e8bdbbf4d24-kube-api-access-wzfcw\") on node \"crc\" DevicePath \"\"" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.811629 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" event={"ID":"2d9dffb1-a917-4fae-b276-4e8bdbbf4d24","Type":"ContainerDied","Data":"52f729a705da6daa4427aa683099630e622ef45f70cded847d5088320593cd86"} Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.811695 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52f729a705da6daa4427aa683099630e622ef45f70cded847d5088320593cd86" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.811701 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fdb222" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.815496 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.815495 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/ac52c329ec8dd388f478b899b6138cb1cc17a7b9bee671d11c6b8bb579k5hzn" event={"ID":"3890f761-4b7f-4033-8397-3b3639ee86a7","Type":"ContainerDied","Data":"dd5f95bfaa6d0c59af4fbbb2b1e3b04044862342fe1cdd55b6c611d158fa51a7"} Dec 11 17:07:59 crc kubenswrapper[5109]: I1211 17:07:59.815719 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd5f95bfaa6d0c59af4fbbb2b1e3b04044862342fe1cdd55b6c611d158fa51a7" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.957393 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/smart-gateway-operator-875f76f9d-7pfqr"] Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958290 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerName="pull" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958306 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerName="pull" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958327 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerName="util" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958334 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerName="util" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958349 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958356 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958376 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958382 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958393 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerName="util" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958399 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerName="util" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958409 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerName="pull" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958415 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerName="pull" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958430 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerName="util" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958443 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerName="util" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958458 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958468 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958486 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerName="pull" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958493 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerName="pull" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958586 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="2d9dffb1-a917-4fae-b276-4e8bdbbf4d24" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958598 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="a88b7761-29b7-4a4c-ab01-951bfec5cad1" containerName="extract" Dec 11 17:08:01 crc kubenswrapper[5109]: I1211 17:08:01.958610 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="3890f761-4b7f-4033-8397-3b3639ee86a7" containerName="extract" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.858229 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.862988 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-operator-dockercfg-j8tzt\"" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.868467 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-875f76f9d-7pfqr"] Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.868564 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp"] Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.873306 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp"] Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.873461 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.875714 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"service-telemetry-operator-dockercfg-qg6t5\"" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.943211 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfd9b\" (UniqueName: \"kubernetes.io/projected/6db9f9d3-3111-4638-85f5-e230c9cf4ccd-kube-api-access-qfd9b\") pod \"service-telemetry-operator-f6b676fbd-rh8fp\" (UID: \"6db9f9d3-3111-4638-85f5-e230c9cf4ccd\") " pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.943275 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/7c3868a0-1147-4baf-a1b7-88ffad30c371-runner\") pod \"smart-gateway-operator-875f76f9d-7pfqr\" (UID: \"7c3868a0-1147-4baf-a1b7-88ffad30c371\") " pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.943302 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6db9f9d3-3111-4638-85f5-e230c9cf4ccd-runner\") pod \"service-telemetry-operator-f6b676fbd-rh8fp\" (UID: \"6db9f9d3-3111-4638-85f5-e230c9cf4ccd\") " pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:03 crc kubenswrapper[5109]: I1211 17:08:03.943327 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm77c\" (UniqueName: \"kubernetes.io/projected/7c3868a0-1147-4baf-a1b7-88ffad30c371-kube-api-access-vm77c\") pod \"smart-gateway-operator-875f76f9d-7pfqr\" (UID: \"7c3868a0-1147-4baf-a1b7-88ffad30c371\") " pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.045081 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-qfd9b\" (UniqueName: \"kubernetes.io/projected/6db9f9d3-3111-4638-85f5-e230c9cf4ccd-kube-api-access-qfd9b\") pod \"service-telemetry-operator-f6b676fbd-rh8fp\" (UID: \"6db9f9d3-3111-4638-85f5-e230c9cf4ccd\") " pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.045168 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/7c3868a0-1147-4baf-a1b7-88ffad30c371-runner\") pod \"smart-gateway-operator-875f76f9d-7pfqr\" (UID: \"7c3868a0-1147-4baf-a1b7-88ffad30c371\") " pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.045231 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6db9f9d3-3111-4638-85f5-e230c9cf4ccd-runner\") pod \"service-telemetry-operator-f6b676fbd-rh8fp\" (UID: \"6db9f9d3-3111-4638-85f5-e230c9cf4ccd\") " pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.045265 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vm77c\" (UniqueName: \"kubernetes.io/projected/7c3868a0-1147-4baf-a1b7-88ffad30c371-kube-api-access-vm77c\") pod \"smart-gateway-operator-875f76f9d-7pfqr\" (UID: \"7c3868a0-1147-4baf-a1b7-88ffad30c371\") " pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.045689 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/7c3868a0-1147-4baf-a1b7-88ffad30c371-runner\") pod \"smart-gateway-operator-875f76f9d-7pfqr\" (UID: \"7c3868a0-1147-4baf-a1b7-88ffad30c371\") " pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.045709 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"runner\" (UniqueName: \"kubernetes.io/empty-dir/6db9f9d3-3111-4638-85f5-e230c9cf4ccd-runner\") pod \"service-telemetry-operator-f6b676fbd-rh8fp\" (UID: \"6db9f9d3-3111-4638-85f5-e230c9cf4ccd\") " pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.067953 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfd9b\" (UniqueName: \"kubernetes.io/projected/6db9f9d3-3111-4638-85f5-e230c9cf4ccd-kube-api-access-qfd9b\") pod \"service-telemetry-operator-f6b676fbd-rh8fp\" (UID: \"6db9f9d3-3111-4638-85f5-e230c9cf4ccd\") " pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.070568 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm77c\" (UniqueName: \"kubernetes.io/projected/7c3868a0-1147-4baf-a1b7-88ffad30c371-kube-api-access-vm77c\") pod \"smart-gateway-operator-875f76f9d-7pfqr\" (UID: \"7c3868a0-1147-4baf-a1b7-88ffad30c371\") " pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.177056 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.190140 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.838235 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/smart-gateway-operator-875f76f9d-7pfqr"] Dec 11 17:08:04 crc kubenswrapper[5109]: I1211 17:08:04.907951 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp"] Dec 11 17:08:04 crc kubenswrapper[5109]: W1211 17:08:04.912641 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6db9f9d3_3111_4638_85f5_e230c9cf4ccd.slice/crio-de14215bb3fa93c8cd05def619c64ee6701b50893c072cd3c7ebbcc46b3dbb52 WatchSource:0}: Error finding container de14215bb3fa93c8cd05def619c64ee6701b50893c072cd3c7ebbcc46b3dbb52: Status 404 returned error can't find the container with id de14215bb3fa93c8cd05def619c64ee6701b50893c072cd3c7ebbcc46b3dbb52 Dec 11 17:08:05 crc kubenswrapper[5109]: I1211 17:08:05.127120 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-8qr9m"] Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.172242 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" event={"ID":"7c3868a0-1147-4baf-a1b7-88ffad30c371","Type":"ContainerStarted","Data":"1194ab9b0ff05ca604b0a4ac61d50cc60578ed8e58f90e0ca0361b4ffd5b7dcd"} Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.172296 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" event={"ID":"6db9f9d3-3111-4638-85f5-e230c9cf4ccd","Type":"ContainerStarted","Data":"de14215bb3fa93c8cd05def619c64ee6701b50893c072cd3c7ebbcc46b3dbb52"} Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.172385 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-8qr9m"] Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.172406 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.174624 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"interconnect-operator-dockercfg-btt5c\"" Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.239110 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnsp4\" (UniqueName: \"kubernetes.io/projected/17398644-29ea-4c2e-befc-e876bf6ecf67-kube-api-access-nnsp4\") pod \"interconnect-operator-78b9bd8798-8qr9m\" (UID: \"17398644-29ea-4c2e-befc-e876bf6ecf67\") " pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.340769 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-nnsp4\" (UniqueName: \"kubernetes.io/projected/17398644-29ea-4c2e-befc-e876bf6ecf67-kube-api-access-nnsp4\") pod \"interconnect-operator-78b9bd8798-8qr9m\" (UID: \"17398644-29ea-4c2e-befc-e876bf6ecf67\") " pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.365932 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnsp4\" (UniqueName: \"kubernetes.io/projected/17398644-29ea-4c2e-befc-e876bf6ecf67-kube-api-access-nnsp4\") pod \"interconnect-operator-78b9bd8798-8qr9m\" (UID: \"17398644-29ea-4c2e-befc-e876bf6ecf67\") " pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.490320 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.716002 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-78b9bd8798-8qr9m"] Dec 11 17:08:06 crc kubenswrapper[5109]: I1211 17:08:06.860259 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" event={"ID":"17398644-29ea-4c2e-befc-e876bf6ecf67","Type":"ContainerStarted","Data":"ceac652dccb99564da8422c71f9cfa7efd98ddf0b454fe0ce190dcbf59485223"} Dec 11 17:08:07 crc kubenswrapper[5109]: I1211 17:08:07.761633 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:08:07 crc kubenswrapper[5109]: I1211 17:08:07.761706 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:08:07 crc kubenswrapper[5109]: I1211 17:08:07.761771 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 17:08:07 crc kubenswrapper[5109]: I1211 17:08:07.762289 5109 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1f3b25744dcfda371142af697057ed29ab8b623200bea5beeda0e00d433c8883"} pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 17:08:07 crc kubenswrapper[5109]: I1211 17:08:07.762603 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" containerID="cri-o://1f3b25744dcfda371142af697057ed29ab8b623200bea5beeda0e00d433c8883" gracePeriod=600 Dec 11 17:08:08 crc kubenswrapper[5109]: I1211 17:08:08.895189 5109 generic.go:358] "Generic (PLEG): container finished" podID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerID="1f3b25744dcfda371142af697057ed29ab8b623200bea5beeda0e00d433c8883" exitCode=0 Dec 11 17:08:08 crc kubenswrapper[5109]: I1211 17:08:08.895247 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerDied","Data":"1f3b25744dcfda371142af697057ed29ab8b623200bea5beeda0e00d433c8883"} Dec 11 17:08:08 crc kubenswrapper[5109]: I1211 17:08:08.895634 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"4c415588c2932db1c71d652e0d9ee8b8cbc78b73b5bf8bdb07993c476779114d"} Dec 11 17:08:08 crc kubenswrapper[5109]: I1211 17:08:08.895658 5109 scope.go:117] "RemoveContainer" containerID="5113406771ea4df98ffaeb4cf5ec00ae40484fd2ce8f067a051e90a693151965" Dec 11 17:08:43 crc kubenswrapper[5109]: I1211 17:08:43.132278 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" event={"ID":"6db9f9d3-3111-4638-85f5-e230c9cf4ccd","Type":"ContainerStarted","Data":"58c5b982d5011a17e388a78e134662bb51758f36bd9e6315b281300642fa7e8d"} Dec 11 17:08:43 crc kubenswrapper[5109]: I1211 17:08:43.133663 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" event={"ID":"7c3868a0-1147-4baf-a1b7-88ffad30c371","Type":"ContainerStarted","Data":"c7a923fe2526368816628f5ef76684f04504269d923f0940058cefe47e291ab6"} Dec 11 17:08:43 crc kubenswrapper[5109]: I1211 17:08:43.134777 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" event={"ID":"17398644-29ea-4c2e-befc-e876bf6ecf67","Type":"ContainerStarted","Data":"2ada01483f466f087255f84a5d235825f680a526620bd05d65a50136b5d4d933"} Dec 11 17:08:43 crc kubenswrapper[5109]: I1211 17:08:43.147701 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/service-telemetry-operator-f6b676fbd-rh8fp" podStartSLOduration=2.363386946 podStartE2EDuration="40.147679745s" podCreationTimestamp="2025-12-11 17:08:03 +0000 UTC" firstStartedPulling="2025-12-11 17:08:04.91562873 +0000 UTC m=+922.595320186" lastFinishedPulling="2025-12-11 17:08:42.699921499 +0000 UTC m=+960.379612985" observedRunningTime="2025-12-11 17:08:43.146149207 +0000 UTC m=+960.825840673" watchObservedRunningTime="2025-12-11 17:08:43.147679745 +0000 UTC m=+960.827371221" Dec 11 17:08:43 crc kubenswrapper[5109]: I1211 17:08:43.165685 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/smart-gateway-operator-875f76f9d-7pfqr" podStartSLOduration=4.379026628 podStartE2EDuration="42.165664134s" podCreationTimestamp="2025-12-11 17:08:01 +0000 UTC" firstStartedPulling="2025-12-11 17:08:04.841037739 +0000 UTC m=+922.520729205" lastFinishedPulling="2025-12-11 17:08:42.627675245 +0000 UTC m=+960.307366711" observedRunningTime="2025-12-11 17:08:43.162145768 +0000 UTC m=+960.841837244" watchObservedRunningTime="2025-12-11 17:08:43.165664134 +0000 UTC m=+960.845355600" Dec 11 17:08:43 crc kubenswrapper[5109]: I1211 17:08:43.188293 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-78b9bd8798-8qr9m" podStartSLOduration=9.775218921 podStartE2EDuration="38.188269246s" podCreationTimestamp="2025-12-11 17:08:05 +0000 UTC" firstStartedPulling="2025-12-11 17:08:06.726018333 +0000 UTC m=+924.405709799" lastFinishedPulling="2025-12-11 17:08:35.139068658 +0000 UTC m=+952.818760124" observedRunningTime="2025-12-11 17:08:43.183680724 +0000 UTC m=+960.863372210" watchObservedRunningTime="2025-12-11 17:08:43.188269246 +0000 UTC m=+960.867960702" Dec 11 17:09:12 crc kubenswrapper[5109]: I1211 17:09:12.411429 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-dqcvs"] Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.422547 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-dqcvs"] Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.423679 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.427997 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-interconnect-sasl-config\"" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.428189 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-users\"" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.428323 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-dockercfg-lfh6r\"" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.428417 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-inter-router-ca\"" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.431035 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-openstack-credentials\"" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.433665 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-inter-router-credentials\"" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.433688 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-openstack-ca\"" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.514845 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.514948 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-config\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.515025 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.515064 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n5sv\" (UniqueName: \"kubernetes.io/projected/fc9beec1-4999-4be9-9dd5-683a7726317e-kube-api-access-6n5sv\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.515100 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-users\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.515286 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.515415 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.617267 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-users\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.617373 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.617415 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.617457 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.617510 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-config\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.617553 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.617593 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-6n5sv\" (UniqueName: \"kubernetes.io/projected/fc9beec1-4999-4be9-9dd5-683a7726317e-kube-api-access-6n5sv\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.620075 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-config\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.623300 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-users\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.623337 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.623350 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.632651 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.634827 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n5sv\" (UniqueName: \"kubernetes.io/projected/fc9beec1-4999-4be9-9dd5-683a7726317e-kube-api-access-6n5sv\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.641065 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-dqcvs\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:13 crc kubenswrapper[5109]: I1211 17:09:13.745159 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:09:14 crc kubenswrapper[5109]: I1211 17:09:14.176048 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-dqcvs"] Dec 11 17:09:14 crc kubenswrapper[5109]: I1211 17:09:14.348143 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" event={"ID":"fc9beec1-4999-4be9-9dd5-683a7726317e","Type":"ContainerStarted","Data":"0f642fc516485d08539e3d0a7cc108e13e0dcc3c52fde47e8509b7b47c7c033f"} Dec 11 17:09:21 crc kubenswrapper[5109]: E1211 17:09:21.892118 5109 certificate_manager.go:613] "Certificate request was not signed" err="timed out waiting for the condition" logger="kubernetes.io/kubelet-serving.UnhandledError" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.016073 5109 certificate_manager.go:566] "Rotating certificates" logger="kubernetes.io/kubelet-serving" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.027601 5109 reflector.go:430] "Caches populated" logger="kubernetes.io/kubelet-serving" type="*v1.CertificateSigningRequest" reflector="k8s.io/client-go/tools/watch/informerwatcher.go:162" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.047318 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58768: no serving certificate available for the kubelet" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.078243 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58784: no serving certificate available for the kubelet" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.124946 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58798: no serving certificate available for the kubelet" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.170907 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58802: no serving certificate available for the kubelet" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.230791 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58804: no serving certificate available for the kubelet" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.330855 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58812: no serving certificate available for the kubelet" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.517806 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58816: no serving certificate available for the kubelet" Dec 11 17:09:24 crc kubenswrapper[5109]: I1211 17:09:24.859206 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58822: no serving certificate available for the kubelet" Dec 11 17:09:25 crc kubenswrapper[5109]: I1211 17:09:25.428251 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" event={"ID":"fc9beec1-4999-4be9-9dd5-683a7726317e","Type":"ContainerStarted","Data":"3484498e803b5859f17664767a8586f5f21a07a2f85b647d75bafb8b3529b1a5"} Dec 11 17:09:25 crc kubenswrapper[5109]: I1211 17:09:25.524587 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58824: no serving certificate available for the kubelet" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.444239 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" podStartSLOduration=4.061428395 podStartE2EDuration="14.444216954s" podCreationTimestamp="2025-12-11 17:09:12 +0000 UTC" firstStartedPulling="2025-12-11 17:09:14.186568893 +0000 UTC m=+991.866260399" lastFinishedPulling="2025-12-11 17:09:24.569357492 +0000 UTC m=+1002.249048958" observedRunningTime="2025-12-11 17:09:25.445559667 +0000 UTC m=+1003.125251163" watchObservedRunningTime="2025-12-11 17:09:26.444216954 +0000 UTC m=+1004.123908420" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.448007 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/prometheus-default-0"] Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.453351 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457042 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"serving-certs-ca-bundle\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457100 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"prometheus-default-rulefiles-0\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457145 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default-tls-assets-0\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457162 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default-web-config\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457224 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-stf-dockercfg-8mwxq\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457226 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-session-secret\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457040 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-prometheus-proxy-tls\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.457257 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"prometheus-default\"" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.462708 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.518823 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/efd0bf2e-ed18-4563-9536-878438477e36-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.518878 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519078 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/efd0bf2e-ed18-4563-9536-878438477e36-tls-assets\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519121 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jfq2b\" (UniqueName: \"kubernetes.io/projected/efd0bf2e-ed18-4563-9536-878438477e36-kube-api-access-jfq2b\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519164 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efd0bf2e-ed18-4563-9536-878438477e36-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519190 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-web-config\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519303 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-config\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519440 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519490 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.519558 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/efd0bf2e-ed18-4563-9536-878438477e36-config-out\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621188 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/efd0bf2e-ed18-4563-9536-878438477e36-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621241 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621283 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/efd0bf2e-ed18-4563-9536-878438477e36-tls-assets\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621300 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jfq2b\" (UniqueName: \"kubernetes.io/projected/efd0bf2e-ed18-4563-9536-878438477e36-kube-api-access-jfq2b\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621316 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efd0bf2e-ed18-4563-9536-878438477e36-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621332 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-web-config\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621363 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-config\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621405 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621430 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.621450 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/efd0bf2e-ed18-4563-9536-878438477e36-config-out\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.622057 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"prometheus-default-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/efd0bf2e-ed18-4563-9536-878438477e36-prometheus-default-rulefiles-0\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: E1211 17:09:26.622200 5109 secret.go:189] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Dec 11 17:09:26 crc kubenswrapper[5109]: E1211 17:09:26.622273 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls podName:efd0bf2e-ed18-4563-9536-878438477e36 nodeName:}" failed. No retries permitted until 2025-12-11 17:09:27.122252655 +0000 UTC m=+1004.801944121 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "efd0bf2e-ed18-4563-9536-878438477e36") : secret "default-prometheus-proxy-tls" not found Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.623134 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efd0bf2e-ed18-4563-9536-878438477e36-configmap-serving-certs-ca-bundle\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.627475 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/efd0bf2e-ed18-4563-9536-878438477e36-config-out\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.627646 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-web-config\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.627980 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-session-secret\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.628125 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-config\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.628481 5109 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.629120 5109 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b68a35fdb49ceb680e0f5fb18658367cc745be4625509f87b4dbf2f4bb7fe0bc/globalmount\"" pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.630984 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/efd0bf2e-ed18-4563-9536-878438477e36-tls-assets\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.639891 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jfq2b\" (UniqueName: \"kubernetes.io/projected/efd0bf2e-ed18-4563-9536-878438477e36-kube-api-access-jfq2b\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.663860 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-f6720e33-c6b5-44b5-b672-115dc2e4a88f\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:26 crc kubenswrapper[5109]: I1211 17:09:26.826716 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58840: no serving certificate available for the kubelet" Dec 11 17:09:27 crc kubenswrapper[5109]: I1211 17:09:27.129969 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:27 crc kubenswrapper[5109]: E1211 17:09:27.130267 5109 secret.go:189] Couldn't get secret service-telemetry/default-prometheus-proxy-tls: secret "default-prometheus-proxy-tls" not found Dec 11 17:09:27 crc kubenswrapper[5109]: E1211 17:09:27.130351 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls podName:efd0bf2e-ed18-4563-9536-878438477e36 nodeName:}" failed. No retries permitted until 2025-12-11 17:09:28.130333063 +0000 UTC m=+1005.810024529 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-prometheus-proxy-tls" (UniqueName: "kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls") pod "prometheus-default-0" (UID: "efd0bf2e-ed18-4563-9536-878438477e36") : secret "default-prometheus-proxy-tls" not found Dec 11 17:09:28 crc kubenswrapper[5109]: I1211 17:09:28.143800 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:28 crc kubenswrapper[5109]: I1211 17:09:28.149012 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-prometheus-proxy-tls\" (UniqueName: \"kubernetes.io/secret/efd0bf2e-ed18-4563-9536-878438477e36-secret-default-prometheus-proxy-tls\") pod \"prometheus-default-0\" (UID: \"efd0bf2e-ed18-4563-9536-878438477e36\") " pod="service-telemetry/prometheus-default-0" Dec 11 17:09:28 crc kubenswrapper[5109]: I1211 17:09:28.276155 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/prometheus-default-0" Dec 11 17:09:28 crc kubenswrapper[5109]: I1211 17:09:28.503940 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/prometheus-default-0"] Dec 11 17:09:29 crc kubenswrapper[5109]: I1211 17:09:29.413402 5109 ???:1] "http: TLS handshake error from 192.168.126.11:58842: no serving certificate available for the kubelet" Dec 11 17:09:29 crc kubenswrapper[5109]: I1211 17:09:29.457639 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"efd0bf2e-ed18-4563-9536-878438477e36","Type":"ContainerStarted","Data":"c1ba2fcbb4390cea30f51a35fe1ca747e85ceaf2fbd2910d6f0fba074b1599d5"} Dec 11 17:09:34 crc kubenswrapper[5109]: I1211 17:09:34.491440 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"efd0bf2e-ed18-4563-9536-878438477e36","Type":"ContainerStarted","Data":"4a842140a18473927a76950ae05a260c277e970da3fb9f759511df47414ee31b"} Dec 11 17:09:34 crc kubenswrapper[5109]: I1211 17:09:34.559138 5109 ???:1] "http: TLS handshake error from 192.168.126.11:48046: no serving certificate available for the kubelet" Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.098707 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb"] Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.135704 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb"] Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.135902 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.269214 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlhwm\" (UniqueName: \"kubernetes.io/projected/8a30f4f4-0565-48b1-93cc-5f5ffc3c131b-kube-api-access-zlhwm\") pod \"default-snmp-webhook-6774d8dfbc-dxnvb\" (UID: \"8a30f4f4-0565-48b1-93cc-5f5ffc3c131b\") " pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.370882 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zlhwm\" (UniqueName: \"kubernetes.io/projected/8a30f4f4-0565-48b1-93cc-5f5ffc3c131b-kube-api-access-zlhwm\") pod \"default-snmp-webhook-6774d8dfbc-dxnvb\" (UID: \"8a30f4f4-0565-48b1-93cc-5f5ffc3c131b\") " pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.391426 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlhwm\" (UniqueName: \"kubernetes.io/projected/8a30f4f4-0565-48b1-93cc-5f5ffc3c131b-kube-api-access-zlhwm\") pod \"default-snmp-webhook-6774d8dfbc-dxnvb\" (UID: \"8a30f4f4-0565-48b1-93cc-5f5ffc3c131b\") " pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.463581 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" Dec 11 17:09:36 crc kubenswrapper[5109]: I1211 17:09:36.714570 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb"] Dec 11 17:09:37 crc kubenswrapper[5109]: I1211 17:09:37.517564 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" event={"ID":"8a30f4f4-0565-48b1-93cc-5f5ffc3c131b","Type":"ContainerStarted","Data":"961d88928079452ad41666b5481eff4984639a4b1c91a703f58969f6782a16dd"} Dec 11 17:09:39 crc kubenswrapper[5109]: I1211 17:09:39.998425 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/alertmanager-default-0"] Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.006627 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.007002 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.008769 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-stf-dockercfg-szmsf\"" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.009021 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-generated\"" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.009144 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-tls-assets-0\"" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.009270 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-cluster-tls-config\"" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.013933 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-default-web-config\"" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.017813 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-alertmanager-proxy-tls\"" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.126823 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.126889 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-config-volume\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.126918 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.126957 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-config-out\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.126982 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.127117 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-tls-assets\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.127203 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zntz5\" (UniqueName: \"kubernetes.io/projected/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-kube-api-access-zntz5\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.127296 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.127345 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-web-config\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228461 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-config-out\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228523 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228550 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-tls-assets\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228572 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-zntz5\" (UniqueName: \"kubernetes.io/projected/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-kube-api-access-zntz5\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228612 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: E1211 17:09:40.228724 5109 secret.go:189] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228811 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-web-config\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: E1211 17:09:40.228820 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls podName:a0e50121-ead0-48f6-940d-3cbbc4b01e5b nodeName:}" failed. No retries permitted until 2025-12-11 17:09:40.728801953 +0000 UTC m=+1018.408493419 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "a0e50121-ead0-48f6-940d-3cbbc4b01e5b") : secret "default-alertmanager-proxy-tls" not found Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228913 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.228998 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-config-volume\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.229052 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.233909 5109 csi_attacher.go:373] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.233975 5109 operation_generator.go:557] "MountVolume.MountDevice succeeded for volume \"pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6046d2242c715eaf186db0616ab849bfde0be4c1ccc2751fc5e6e7260e86794f/globalmount\"" pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.235680 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"cluster-tls-config\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-cluster-tls-config\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.236599 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-web-config\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.237221 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-tls-assets\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.238320 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-config-out\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.240506 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-config-volume\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.242140 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-session-secret\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-session-secret\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.254858 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-zntz5\" (UniqueName: \"kubernetes.io/projected/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-kube-api-access-zntz5\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.276982 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6688562c-3020-42ec-8f4f-f0077f76ab62\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: I1211 17:09:40.738285 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:40 crc kubenswrapper[5109]: E1211 17:09:40.738455 5109 secret.go:189] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Dec 11 17:09:40 crc kubenswrapper[5109]: E1211 17:09:40.738729 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls podName:a0e50121-ead0-48f6-940d-3cbbc4b01e5b nodeName:}" failed. No retries permitted until 2025-12-11 17:09:41.738703815 +0000 UTC m=+1019.418395281 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "a0e50121-ead0-48f6-940d-3cbbc4b01e5b") : secret "default-alertmanager-proxy-tls" not found Dec 11 17:09:41 crc kubenswrapper[5109]: I1211 17:09:41.568581 5109 generic.go:358] "Generic (PLEG): container finished" podID="efd0bf2e-ed18-4563-9536-878438477e36" containerID="4a842140a18473927a76950ae05a260c277e970da3fb9f759511df47414ee31b" exitCode=0 Dec 11 17:09:41 crc kubenswrapper[5109]: I1211 17:09:41.568688 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"efd0bf2e-ed18-4563-9536-878438477e36","Type":"ContainerDied","Data":"4a842140a18473927a76950ae05a260c277e970da3fb9f759511df47414ee31b"} Dec 11 17:09:41 crc kubenswrapper[5109]: I1211 17:09:41.569839 5109 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 17:09:41 crc kubenswrapper[5109]: I1211 17:09:41.754203 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:41 crc kubenswrapper[5109]: E1211 17:09:41.754376 5109 secret.go:189] Couldn't get secret service-telemetry/default-alertmanager-proxy-tls: secret "default-alertmanager-proxy-tls" not found Dec 11 17:09:41 crc kubenswrapper[5109]: E1211 17:09:41.755083 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls podName:a0e50121-ead0-48f6-940d-3cbbc4b01e5b nodeName:}" failed. No retries permitted until 2025-12-11 17:09:43.755052145 +0000 UTC m=+1021.434743801 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "secret-default-alertmanager-proxy-tls" (UniqueName: "kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls") pod "alertmanager-default-0" (UID: "a0e50121-ead0-48f6-940d-3cbbc4b01e5b") : secret "default-alertmanager-proxy-tls" not found Dec 11 17:09:43 crc kubenswrapper[5109]: I1211 17:09:43.793207 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:43 crc kubenswrapper[5109]: I1211 17:09:43.803618 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-default-alertmanager-proxy-tls\" (UniqueName: \"kubernetes.io/secret/a0e50121-ead0-48f6-940d-3cbbc4b01e5b-secret-default-alertmanager-proxy-tls\") pod \"alertmanager-default-0\" (UID: \"a0e50121-ead0-48f6-940d-3cbbc4b01e5b\") " pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:43 crc kubenswrapper[5109]: I1211 17:09:43.941073 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"alertmanager-stf-dockercfg-szmsf\"" Dec 11 17:09:43 crc kubenswrapper[5109]: I1211 17:09:43.950066 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/alertmanager-default-0" Dec 11 17:09:44 crc kubenswrapper[5109]: I1211 17:09:44.824266 5109 ???:1] "http: TLS handshake error from 192.168.126.11:39280: no serving certificate available for the kubelet" Dec 11 17:09:46 crc kubenswrapper[5109]: I1211 17:09:46.659505 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/alertmanager-default-0"] Dec 11 17:09:47 crc kubenswrapper[5109]: I1211 17:09:47.610243 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" event={"ID":"8a30f4f4-0565-48b1-93cc-5f5ffc3c131b","Type":"ContainerStarted","Data":"8518ed06dad27b3365f8cefc80abe3c6bd362beae671a08f9ead6e61e9f1d179"} Dec 11 17:09:47 crc kubenswrapper[5109]: I1211 17:09:47.611924 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a0e50121-ead0-48f6-940d-3cbbc4b01e5b","Type":"ContainerStarted","Data":"5e8019aa68155d26c243c009576a34d868519e8aa97e7d12187254ad9e67b372"} Dec 11 17:09:47 crc kubenswrapper[5109]: I1211 17:09:47.626527 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-snmp-webhook-6774d8dfbc-dxnvb" podStartSLOduration=1.610269795 podStartE2EDuration="11.626510134s" podCreationTimestamp="2025-12-11 17:09:36 +0000 UTC" firstStartedPulling="2025-12-11 17:09:36.715585719 +0000 UTC m=+1014.395277225" lastFinishedPulling="2025-12-11 17:09:46.731826098 +0000 UTC m=+1024.411517564" observedRunningTime="2025-12-11 17:09:47.622156208 +0000 UTC m=+1025.301847674" watchObservedRunningTime="2025-12-11 17:09:47.626510134 +0000 UTC m=+1025.306201600" Dec 11 17:09:48 crc kubenswrapper[5109]: I1211 17:09:48.620043 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a0e50121-ead0-48f6-940d-3cbbc4b01e5b","Type":"ContainerStarted","Data":"0760bd9580acced118594a69d825f340dadae508a29d619fb00340bee074c873"} Dec 11 17:09:51 crc kubenswrapper[5109]: I1211 17:09:51.650595 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"efd0bf2e-ed18-4563-9536-878438477e36","Type":"ContainerStarted","Data":"660c710a7019387847cf9bf51c5af8475a9a488fa75eb13de220efbb578cbc1d"} Dec 11 17:09:52 crc kubenswrapper[5109]: I1211 17:09:52.658420 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"efd0bf2e-ed18-4563-9536-878438477e36","Type":"ContainerStarted","Data":"d0ac6c36e8eb74ac58acb121c305e36c0bc2cd9d6aa9e1513b8ece5f580d5b6d"} Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.159707 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9"] Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.168249 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.170578 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-meter-proxy-tls\"" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.170929 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9"] Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.170945 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-session-secret\"" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.171396 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"smart-gateway-dockercfg-xzwtg\"" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.171949 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-meter-sg-core-configmap\"" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.223862 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75jbj\" (UniqueName: \"kubernetes.io/projected/d251c432-204d-4df1-a193-fb361eba6102-kube-api-access-75jbj\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.223920 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.223949 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/d251c432-204d-4df1-a193-fb361eba6102-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.224013 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.224205 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/d251c432-204d-4df1-a193-fb361eba6102-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.326093 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.326191 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/d251c432-204d-4df1-a193-fb361eba6102-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.326289 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-75jbj\" (UniqueName: \"kubernetes.io/projected/d251c432-204d-4df1-a193-fb361eba6102-kube-api-access-75jbj\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.326330 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.326360 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/d251c432-204d-4df1-a193-fb361eba6102-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: E1211 17:09:53.326756 5109 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Dec 11 17:09:53 crc kubenswrapper[5109]: E1211 17:09:53.326854 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls podName:d251c432-204d-4df1-a193-fb361eba6102 nodeName:}" failed. No retries permitted until 2025-12-11 17:09:53.826830633 +0000 UTC m=+1031.506522099 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-787645d794-45cc9" (UID: "d251c432-204d-4df1-a193-fb361eba6102") : secret "default-cloud1-coll-meter-proxy-tls" not found Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.327200 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/d251c432-204d-4df1-a193-fb361eba6102-socket-dir\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.327316 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/d251c432-204d-4df1-a193-fb361eba6102-sg-core-config\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.334661 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-session-secret\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.347546 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-75jbj\" (UniqueName: \"kubernetes.io/projected/d251c432-204d-4df1-a193-fb361eba6102-kube-api-access-75jbj\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: I1211 17:09:53.834313 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:53 crc kubenswrapper[5109]: E1211 17:09:53.834497 5109 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-coll-meter-proxy-tls: secret "default-cloud1-coll-meter-proxy-tls" not found Dec 11 17:09:53 crc kubenswrapper[5109]: E1211 17:09:53.834702 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls podName:d251c432-204d-4df1-a193-fb361eba6102 nodeName:}" failed. No retries permitted until 2025-12-11 17:09:54.834682744 +0000 UTC m=+1032.514374210 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-coll-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls") pod "default-cloud1-coll-meter-smartgateway-787645d794-45cc9" (UID: "d251c432-204d-4df1-a193-fb361eba6102") : secret "default-cloud1-coll-meter-proxy-tls" not found Dec 11 17:09:54 crc kubenswrapper[5109]: I1211 17:09:54.671813 5109 generic.go:358] "Generic (PLEG): container finished" podID="a0e50121-ead0-48f6-940d-3cbbc4b01e5b" containerID="0760bd9580acced118594a69d825f340dadae508a29d619fb00340bee074c873" exitCode=0 Dec 11 17:09:54 crc kubenswrapper[5109]: I1211 17:09:54.671894 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a0e50121-ead0-48f6-940d-3cbbc4b01e5b","Type":"ContainerDied","Data":"0760bd9580acced118594a69d825f340dadae508a29d619fb00340bee074c873"} Dec 11 17:09:54 crc kubenswrapper[5109]: I1211 17:09:54.848360 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:54 crc kubenswrapper[5109]: I1211 17:09:54.852827 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-coll-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/d251c432-204d-4df1-a193-fb361eba6102-default-cloud1-coll-meter-proxy-tls\") pod \"default-cloud1-coll-meter-smartgateway-787645d794-45cc9\" (UID: \"d251c432-204d-4df1-a193-fb361eba6102\") " pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:54 crc kubenswrapper[5109]: I1211 17:09:54.986020 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.529344 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65"] Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.939531 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65"] Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.939725 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.941656 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-meter-proxy-tls\"" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.941678 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-meter-sg-core-configmap\"" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.963107 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.963151 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.963193 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.963214 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:55 crc kubenswrapper[5109]: I1211 17:09:55.963263 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx5lr\" (UniqueName: \"kubernetes.io/projected/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-kube-api-access-vx5lr\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.066553 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.066612 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.066670 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.066704 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.066768 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vx5lr\" (UniqueName: \"kubernetes.io/projected/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-kube-api-access-vx5lr\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: E1211 17:09:56.067549 5109 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 11 17:09:56 crc kubenswrapper[5109]: E1211 17:09:56.067723 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls podName:e0ad6477-4b84-4aae-9b44-7ca1cbc17690 nodeName:}" failed. No retries permitted until 2025-12-11 17:09:56.567634218 +0000 UTC m=+1034.247325694 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" (UID: "e0ad6477-4b84-4aae-9b44-7ca1cbc17690") : secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.068519 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-socket-dir\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.069186 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-sg-core-config\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.088390 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx5lr\" (UniqueName: \"kubernetes.io/projected/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-kube-api-access-vx5lr\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.092327 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-session-secret\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: I1211 17:09:56.574759 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:56 crc kubenswrapper[5109]: E1211 17:09:56.574944 5109 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-ceil-meter-proxy-tls: secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 11 17:09:56 crc kubenswrapper[5109]: E1211 17:09:56.575023 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls podName:e0ad6477-4b84-4aae-9b44-7ca1cbc17690 nodeName:}" failed. No retries permitted until 2025-12-11 17:09:57.575003659 +0000 UTC m=+1035.254695115 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-ceil-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls") pod "default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" (UID: "e0ad6477-4b84-4aae-9b44-7ca1cbc17690") : secret "default-cloud1-ceil-meter-proxy-tls" not found Dec 11 17:09:57 crc kubenswrapper[5109]: I1211 17:09:57.590553 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:57 crc kubenswrapper[5109]: I1211 17:09:57.595676 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-ceil-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/e0ad6477-4b84-4aae-9b44-7ca1cbc17690-default-cloud1-ceil-meter-proxy-tls\") pod \"default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65\" (UID: \"e0ad6477-4b84-4aae-9b44-7ca1cbc17690\") " pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:09:57 crc kubenswrapper[5109]: I1211 17:09:57.762833 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" Dec 11 17:10:01 crc kubenswrapper[5109]: I1211 17:10:01.690606 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56"] Dec 11 17:10:01 crc kubenswrapper[5109]: I1211 17:10:01.953337 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56"] Dec 11 17:10:01 crc kubenswrapper[5109]: I1211 17:10:01.954097 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:01 crc kubenswrapper[5109]: I1211 17:10:01.957010 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-sens-meter-sg-core-configmap\"" Dec 11 17:10:01 crc kubenswrapper[5109]: I1211 17:10:01.957383 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-cloud1-sens-meter-proxy-tls\"" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.057846 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/80337a91-4532-4bc3-b340-39ba1681f3a1-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.057989 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/80337a91-4532-4bc3-b340-39ba1681f3a1-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.058127 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.058275 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzd6k\" (UniqueName: \"kubernetes.io/projected/80337a91-4532-4bc3-b340-39ba1681f3a1-kube-api-access-hzd6k\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.058454 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.159604 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hzd6k\" (UniqueName: \"kubernetes.io/projected/80337a91-4532-4bc3-b340-39ba1681f3a1-kube-api-access-hzd6k\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.159667 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.159704 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/80337a91-4532-4bc3-b340-39ba1681f3a1-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.159766 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/80337a91-4532-4bc3-b340-39ba1681f3a1-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.159782 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: E1211 17:10:02.160699 5109 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Dec 11 17:10:02 crc kubenswrapper[5109]: E1211 17:10:02.160998 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls podName:80337a91-4532-4bc3-b340-39ba1681f3a1 nodeName:}" failed. No retries permitted until 2025-12-11 17:10:02.660977801 +0000 UTC m=+1040.340669267 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" (UID: "80337a91-4532-4bc3-b340-39ba1681f3a1") : secret "default-cloud1-sens-meter-proxy-tls" not found Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.161217 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/80337a91-4532-4bc3-b340-39ba1681f3a1-socket-dir\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.161707 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/80337a91-4532-4bc3-b340-39ba1681f3a1-sg-core-config\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.166277 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"session-secret\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-session-secret\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.177766 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzd6k\" (UniqueName: \"kubernetes.io/projected/80337a91-4532-4bc3-b340-39ba1681f3a1-kube-api-access-hzd6k\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: I1211 17:10:02.666281 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:02 crc kubenswrapper[5109]: E1211 17:10:02.666517 5109 secret.go:189] Couldn't get secret service-telemetry/default-cloud1-sens-meter-proxy-tls: secret "default-cloud1-sens-meter-proxy-tls" not found Dec 11 17:10:02 crc kubenswrapper[5109]: E1211 17:10:02.666587 5109 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls podName:80337a91-4532-4bc3-b340-39ba1681f3a1 nodeName:}" failed. No retries permitted until 2025-12-11 17:10:03.666566848 +0000 UTC m=+1041.346258314 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "default-cloud1-sens-meter-proxy-tls" (UniqueName: "kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls") pod "default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" (UID: "80337a91-4532-4bc3-b340-39ba1681f3a1") : secret "default-cloud1-sens-meter-proxy-tls" not found Dec 11 17:10:03 crc kubenswrapper[5109]: I1211 17:10:03.679807 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:03 crc kubenswrapper[5109]: I1211 17:10:03.698490 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-cloud1-sens-meter-proxy-tls\" (UniqueName: \"kubernetes.io/secret/80337a91-4532-4bc3-b340-39ba1681f3a1-default-cloud1-sens-meter-proxy-tls\") pod \"default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56\" (UID: \"80337a91-4532-4bc3-b340-39ba1681f3a1\") " pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:03 crc kubenswrapper[5109]: I1211 17:10:03.774173 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" Dec 11 17:10:04 crc kubenswrapper[5109]: I1211 17:10:04.479650 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65"] Dec 11 17:10:04 crc kubenswrapper[5109]: I1211 17:10:04.569134 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9"] Dec 11 17:10:04 crc kubenswrapper[5109]: I1211 17:10:04.581568 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56"] Dec 11 17:10:05 crc kubenswrapper[5109]: W1211 17:10:05.098438 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80337a91_4532_4bc3_b340_39ba1681f3a1.slice/crio-099f51379d95d1b7989ee2d581f129e827bddd850098d2edc8e6372098d19cd5 WatchSource:0}: Error finding container 099f51379d95d1b7989ee2d581f129e827bddd850098d2edc8e6372098d19cd5: Status 404 returned error can't find the container with id 099f51379d95d1b7989ee2d581f129e827bddd850098d2edc8e6372098d19cd5 Dec 11 17:10:05 crc kubenswrapper[5109]: W1211 17:10:05.102032 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd251c432_204d_4df1_a193_fb361eba6102.slice/crio-e263e8120eb86184e44a6510d2fcb6408494ea9cf64c3eb9d87b3d5ab0adee96 WatchSource:0}: Error finding container e263e8120eb86184e44a6510d2fcb6408494ea9cf64c3eb9d87b3d5ab0adee96: Status 404 returned error can't find the container with id e263e8120eb86184e44a6510d2fcb6408494ea9cf64c3eb9d87b3d5ab0adee96 Dec 11 17:10:05 crc kubenswrapper[5109]: I1211 17:10:05.327359 5109 ???:1] "http: TLS handshake error from 192.168.126.11:50356: no serving certificate available for the kubelet" Dec 11 17:10:05 crc kubenswrapper[5109]: I1211 17:10:05.754217 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a0e50121-ead0-48f6-940d-3cbbc4b01e5b","Type":"ContainerStarted","Data":"0115787e70e268f437d4596e84bb656e6a3dbddcdfd4da603197faf797c8079a"} Dec 11 17:10:05 crc kubenswrapper[5109]: I1211 17:10:05.755468 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerStarted","Data":"e263e8120eb86184e44a6510d2fcb6408494ea9cf64c3eb9d87b3d5ab0adee96"} Dec 11 17:10:05 crc kubenswrapper[5109]: I1211 17:10:05.756793 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerStarted","Data":"ec3d7cd9e168459cf3132a6ade96bc64eb4b6b6efc614a4f2210bd9fff3ca789"} Dec 11 17:10:05 crc kubenswrapper[5109]: I1211 17:10:05.757726 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerStarted","Data":"099f51379d95d1b7989ee2d581f129e827bddd850098d2edc8e6372098d19cd5"} Dec 11 17:10:05 crc kubenswrapper[5109]: I1211 17:10:05.761381 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/prometheus-default-0" event={"ID":"efd0bf2e-ed18-4563-9536-878438477e36","Type":"ContainerStarted","Data":"a95b599786616c8fb607d75b06a7d6e30b2fe759e4259d0729d462d02b516c06"} Dec 11 17:10:05 crc kubenswrapper[5109]: I1211 17:10:05.786418 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/prometheus-default-0" podStartSLOduration=4.071287446 podStartE2EDuration="40.78639492s" podCreationTimestamp="2025-12-11 17:09:25 +0000 UTC" firstStartedPulling="2025-12-11 17:09:28.522568729 +0000 UTC m=+1006.202260195" lastFinishedPulling="2025-12-11 17:10:05.237676213 +0000 UTC m=+1042.917367669" observedRunningTime="2025-12-11 17:10:05.784281957 +0000 UTC m=+1043.463973443" watchObservedRunningTime="2025-12-11 17:10:05.78639492 +0000 UTC m=+1043.466086386" Dec 11 17:10:06 crc kubenswrapper[5109]: I1211 17:10:06.770252 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerStarted","Data":"25a14c32125034eb2ccbdb8921f427688a4ba5e1bfff8b7205d43a887db292ee"} Dec 11 17:10:06 crc kubenswrapper[5109]: I1211 17:10:06.771867 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerStarted","Data":"4b529f97b593909b320961c986d6a7bfbbef3928cc68e49c0cedb515503b4d6a"} Dec 11 17:10:06 crc kubenswrapper[5109]: I1211 17:10:06.774216 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerStarted","Data":"a4942fe6e9a0d1548d404bbf7ebdee8f5c92d5784c4a17f52cb08a0ad8fc49c3"} Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.699215 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj"] Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.706885 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.712110 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-coll-event-sg-core-configmap\"" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.712407 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"elasticsearch-es-cert\"" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.714251 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj"] Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.794940 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a0e50121-ead0-48f6-940d-3cbbc4b01e5b","Type":"ContainerStarted","Data":"e27303a7fb53d95875ed3826d96bd48b98148ca0521040cdf7c36e671cb93566"} Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.846256 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzsb6\" (UniqueName: \"kubernetes.io/projected/cbadc371-418d-4601-90c4-1d5d3bd787e7-kube-api-access-xzsb6\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.846465 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cbadc371-418d-4601-90c4-1d5d3bd787e7-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.846589 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cbadc371-418d-4601-90c4-1d5d3bd787e7-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.846620 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/cbadc371-418d-4601-90c4-1d5d3bd787e7-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.947621 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cbadc371-418d-4601-90c4-1d5d3bd787e7-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.947683 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/cbadc371-418d-4601-90c4-1d5d3bd787e7-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.947837 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xzsb6\" (UniqueName: \"kubernetes.io/projected/cbadc371-418d-4601-90c4-1d5d3bd787e7-kube-api-access-xzsb6\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.947961 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cbadc371-418d-4601-90c4-1d5d3bd787e7-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.948087 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/cbadc371-418d-4601-90c4-1d5d3bd787e7-socket-dir\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.949342 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/cbadc371-418d-4601-90c4-1d5d3bd787e7-sg-core-config\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.954974 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/cbadc371-418d-4601-90c4-1d5d3bd787e7-elastic-certs\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:07 crc kubenswrapper[5109]: I1211 17:10:07.964283 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzsb6\" (UniqueName: \"kubernetes.io/projected/cbadc371-418d-4601-90c4-1d5d3bd787e7-kube-api-access-xzsb6\") pod \"default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj\" (UID: \"cbadc371-418d-4601-90c4-1d5d3bd787e7\") " pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:08 crc kubenswrapper[5109]: I1211 17:10:08.128360 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" Dec 11 17:10:08 crc kubenswrapper[5109]: I1211 17:10:08.277200 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/prometheus-default-0" Dec 11 17:10:08 crc kubenswrapper[5109]: I1211 17:10:08.582982 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj"] Dec 11 17:10:08 crc kubenswrapper[5109]: W1211 17:10:08.592972 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbadc371_418d_4601_90c4_1d5d3bd787e7.slice/crio-78775c66f75aa7c3d23b085ca0f8457d048928d4773f12041fe982a3801a0126 WatchSource:0}: Error finding container 78775c66f75aa7c3d23b085ca0f8457d048928d4773f12041fe982a3801a0126: Status 404 returned error can't find the container with id 78775c66f75aa7c3d23b085ca0f8457d048928d4773f12041fe982a3801a0126 Dec 11 17:10:08 crc kubenswrapper[5109]: I1211 17:10:08.807224 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" event={"ID":"cbadc371-418d-4601-90c4-1d5d3bd787e7","Type":"ContainerStarted","Data":"78775c66f75aa7c3d23b085ca0f8457d048928d4773f12041fe982a3801a0126"} Dec 11 17:10:08 crc kubenswrapper[5109]: I1211 17:10:08.810430 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/alertmanager-default-0" event={"ID":"a0e50121-ead0-48f6-940d-3cbbc4b01e5b","Type":"ContainerStarted","Data":"1e02e01bc16a571d88717482460b70668dae831553133c92f1848744fcaee47d"} Dec 11 17:10:08 crc kubenswrapper[5109]: I1211 17:10:08.839031 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/alertmanager-default-0" podStartSLOduration=17.804343 podStartE2EDuration="30.83901322s" podCreationTimestamp="2025-12-11 17:09:38 +0000 UTC" firstStartedPulling="2025-12-11 17:09:54.673167577 +0000 UTC m=+1032.352859043" lastFinishedPulling="2025-12-11 17:10:07.707837807 +0000 UTC m=+1045.387529263" observedRunningTime="2025-12-11 17:10:08.835318149 +0000 UTC m=+1046.515009625" watchObservedRunningTime="2025-12-11 17:10:08.83901322 +0000 UTC m=+1046.518704686" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.298073 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc"] Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.311032 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.313250 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"default-cloud1-ceil-event-sg-core-configmap\"" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.314046 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc"] Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.389626 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/250da305-9832-4e46-9a23-7edcfa9d6682-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.389721 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/250da305-9832-4e46-9a23-7edcfa9d6682-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.389802 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/250da305-9832-4e46-9a23-7edcfa9d6682-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.389969 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9c6v\" (UniqueName: \"kubernetes.io/projected/250da305-9832-4e46-9a23-7edcfa9d6682-kube-api-access-v9c6v\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.491675 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/250da305-9832-4e46-9a23-7edcfa9d6682-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.492703 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/250da305-9832-4e46-9a23-7edcfa9d6682-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.492622 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sg-core-config\" (UniqueName: \"kubernetes.io/configmap/250da305-9832-4e46-9a23-7edcfa9d6682-sg-core-config\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.493622 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/250da305-9832-4e46-9a23-7edcfa9d6682-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.493882 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-v9c6v\" (UniqueName: \"kubernetes.io/projected/250da305-9832-4e46-9a23-7edcfa9d6682-kube-api-access-v9c6v\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.493977 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/empty-dir/250da305-9832-4e46-9a23-7edcfa9d6682-socket-dir\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.499614 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"elastic-certs\" (UniqueName: \"kubernetes.io/secret/250da305-9832-4e46-9a23-7edcfa9d6682-elastic-certs\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.517204 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9c6v\" (UniqueName: \"kubernetes.io/projected/250da305-9832-4e46-9a23-7edcfa9d6682-kube-api-access-v9c6v\") pod \"default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc\" (UID: \"250da305-9832-4e46-9a23-7edcfa9d6682\") " pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:09 crc kubenswrapper[5109]: I1211 17:10:09.637013 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.276489 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/prometheus-default-0" Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.309846 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/prometheus-default-0" Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.844642 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc"] Dec 11 17:10:13 crc kubenswrapper[5109]: W1211 17:10:13.850443 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod250da305_9832_4e46_9a23_7edcfa9d6682.slice/crio-5e5f2f64fb61e6613c520c811dd5f4dd3b853e9789c775f9d3ddd14468d5d089 WatchSource:0}: Error finding container 5e5f2f64fb61e6613c520c811dd5f4dd3b853e9789c775f9d3ddd14468d5d089: Status 404 returned error can't find the container with id 5e5f2f64fb61e6613c520c811dd5f4dd3b853e9789c775f9d3ddd14468d5d089 Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.854102 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerStarted","Data":"b65fce5b231a06dcfc337ca08c0d58faad359207a2dd28028bc4047249326b99"} Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.857912 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerStarted","Data":"4d45043e287b4ba0050304015feca494fc9d187776b5432dff83cd00b0e7ea95"} Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.859370 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" event={"ID":"cbadc371-418d-4601-90c4-1d5d3bd787e7","Type":"ContainerStarted","Data":"6e689757652cfc3813f922639c3fdb06df5d0c4b5da6d36f1aa306bcdda9e732"} Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.861517 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerStarted","Data":"3d77ad790743ed319d5abf5bd0bf2a2dc023bc6ba6e624f2e453d8f3808e0ea3"} Dec 11 17:10:13 crc kubenswrapper[5109]: I1211 17:10:13.895812 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/prometheus-default-0" Dec 11 17:10:14 crc kubenswrapper[5109]: I1211 17:10:14.872704 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" event={"ID":"250da305-9832-4e46-9a23-7edcfa9d6682","Type":"ContainerStarted","Data":"5e5f2f64fb61e6613c520c811dd5f4dd3b853e9789c775f9d3ddd14468d5d089"} Dec 11 17:10:16 crc kubenswrapper[5109]: I1211 17:10:16.886925 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" event={"ID":"250da305-9832-4e46-9a23-7edcfa9d6682","Type":"ContainerStarted","Data":"9b00465312c11878d4bb79e4c15cbf622f48f2d0fae2c0f6c39d3e1904433be8"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.390433 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-dqcvs"] Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.391162 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" podUID="fc9beec1-4999-4be9-9dd5-683a7726317e" containerName="default-interconnect" containerID="cri-o://3484498e803b5859f17664767a8586f5f21a07a2f85b647d75bafb8b3529b1a5" gracePeriod=30 Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.932092 5109 generic.go:358] "Generic (PLEG): container finished" podID="fc9beec1-4999-4be9-9dd5-683a7726317e" containerID="3484498e803b5859f17664767a8586f5f21a07a2f85b647d75bafb8b3529b1a5" exitCode=0 Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.932264 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" event={"ID":"fc9beec1-4999-4be9-9dd5-683a7726317e","Type":"ContainerDied","Data":"3484498e803b5859f17664767a8586f5f21a07a2f85b647d75bafb8b3529b1a5"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.934787 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerStarted","Data":"f9473b96632837cf365824d24ad7d6635a4d25ea3e910791f4ff00259df9c992"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.937991 5109 generic.go:358] "Generic (PLEG): container finished" podID="e0ad6477-4b84-4aae-9b44-7ca1cbc17690" containerID="4d45043e287b4ba0050304015feca494fc9d187776b5432dff83cd00b0e7ea95" exitCode=0 Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.938038 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerDied","Data":"4d45043e287b4ba0050304015feca494fc9d187776b5432dff83cd00b0e7ea95"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.938053 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerStarted","Data":"d003fa66c960713be1f9d603decd6e1dcaeeaea624d04980c4f6b8622b0dcf18"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.938316 5109 scope.go:117] "RemoveContainer" containerID="4d45043e287b4ba0050304015feca494fc9d187776b5432dff83cd00b0e7ea95" Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.943962 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" event={"ID":"cbadc371-418d-4601-90c4-1d5d3bd787e7","Type":"ContainerStarted","Data":"b3cb54ba8bfc849ff9f8e841b36406c4075c9a45f4813e36059eb189caa712cb"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.946550 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerStarted","Data":"9a478cf8b3529916272320211ee6658d804e411a25368c7b2c50697d63b31633"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.948553 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" event={"ID":"250da305-9832-4e46-9a23-7edcfa9d6682","Type":"ContainerStarted","Data":"4e8a3f18b59efd9fb380ca5409180976ce92d0adaad111898cd845136ea1130e"} Dec 11 17:10:21 crc kubenswrapper[5109]: I1211 17:10:21.963575 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" podStartSLOduration=13.287795532 podStartE2EDuration="28.963560507s" podCreationTimestamp="2025-12-11 17:09:53 +0000 UTC" firstStartedPulling="2025-12-11 17:10:05.104353495 +0000 UTC m=+1042.784044961" lastFinishedPulling="2025-12-11 17:10:20.78011846 +0000 UTC m=+1058.459809936" observedRunningTime="2025-12-11 17:10:21.958429102 +0000 UTC m=+1059.638120568" watchObservedRunningTime="2025-12-11 17:10:21.963560507 +0000 UTC m=+1059.643251963" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.002864 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" podStartSLOduration=2.93428666 podStartE2EDuration="15.002842164s" podCreationTimestamp="2025-12-11 17:10:07 +0000 UTC" firstStartedPulling="2025-12-11 17:10:08.598848966 +0000 UTC m=+1046.278540432" lastFinishedPulling="2025-12-11 17:10:20.66740447 +0000 UTC m=+1058.347095936" observedRunningTime="2025-12-11 17:10:21.999538022 +0000 UTC m=+1059.679229498" watchObservedRunningTime="2025-12-11 17:10:22.002842164 +0000 UTC m=+1059.682533630" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.020756 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" podStartSLOduration=6.020761751 podStartE2EDuration="13.020723072s" podCreationTimestamp="2025-12-11 17:10:09 +0000 UTC" firstStartedPulling="2025-12-11 17:10:13.851879661 +0000 UTC m=+1051.531571127" lastFinishedPulling="2025-12-11 17:10:20.851840982 +0000 UTC m=+1058.531532448" observedRunningTime="2025-12-11 17:10:22.016010357 +0000 UTC m=+1059.695701853" watchObservedRunningTime="2025-12-11 17:10:22.020723072 +0000 UTC m=+1059.700414538" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.036905 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" podStartSLOduration=5.391102432 podStartE2EDuration="21.03687952s" podCreationTimestamp="2025-12-11 17:10:01 +0000 UTC" firstStartedPulling="2025-12-11 17:10:05.099957397 +0000 UTC m=+1042.779648863" lastFinishedPulling="2025-12-11 17:10:20.745734485 +0000 UTC m=+1058.425425951" observedRunningTime="2025-12-11 17:10:22.031475757 +0000 UTC m=+1059.711167233" watchObservedRunningTime="2025-12-11 17:10:22.03687952 +0000 UTC m=+1059.716570986" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.311121 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.343672 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-m97qq"] Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.344822 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="fc9beec1-4999-4be9-9dd5-683a7726317e" containerName="default-interconnect" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.344918 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc9beec1-4999-4be9-9dd5-683a7726317e" containerName="default-interconnect" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.345162 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="fc9beec1-4999-4be9-9dd5-683a7726317e" containerName="default-interconnect" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.493284 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-credentials\") pod \"fc9beec1-4999-4be9-9dd5-683a7726317e\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.493419 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-config\") pod \"fc9beec1-4999-4be9-9dd5-683a7726317e\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.493472 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-users\") pod \"fc9beec1-4999-4be9-9dd5-683a7726317e\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.493513 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-ca\") pod \"fc9beec1-4999-4be9-9dd5-683a7726317e\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.493544 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n5sv\" (UniqueName: \"kubernetes.io/projected/fc9beec1-4999-4be9-9dd5-683a7726317e-kube-api-access-6n5sv\") pod \"fc9beec1-4999-4be9-9dd5-683a7726317e\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.493570 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-ca\") pod \"fc9beec1-4999-4be9-9dd5-683a7726317e\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.493628 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-credentials\") pod \"fc9beec1-4999-4be9-9dd5-683a7726317e\" (UID: \"fc9beec1-4999-4be9-9dd5-683a7726317e\") " Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.499442 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-config" (OuterVolumeSpecName: "sasl-config") pod "fc9beec1-4999-4be9-9dd5-683a7726317e" (UID: "fc9beec1-4999-4be9-9dd5-683a7726317e"). InnerVolumeSpecName "sasl-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.503899 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-ca" (OuterVolumeSpecName: "default-interconnect-openstack-ca") pod "fc9beec1-4999-4be9-9dd5-683a7726317e" (UID: "fc9beec1-4999-4be9-9dd5-683a7726317e"). InnerVolumeSpecName "default-interconnect-openstack-ca". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.503944 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-ca" (OuterVolumeSpecName: "default-interconnect-inter-router-ca") pod "fc9beec1-4999-4be9-9dd5-683a7726317e" (UID: "fc9beec1-4999-4be9-9dd5-683a7726317e"). InnerVolumeSpecName "default-interconnect-inter-router-ca". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.504227 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-credentials" (OuterVolumeSpecName: "default-interconnect-openstack-credentials") pod "fc9beec1-4999-4be9-9dd5-683a7726317e" (UID: "fc9beec1-4999-4be9-9dd5-683a7726317e"). InnerVolumeSpecName "default-interconnect-openstack-credentials". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.504299 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc9beec1-4999-4be9-9dd5-683a7726317e-kube-api-access-6n5sv" (OuterVolumeSpecName: "kube-api-access-6n5sv") pod "fc9beec1-4999-4be9-9dd5-683a7726317e" (UID: "fc9beec1-4999-4be9-9dd5-683a7726317e"). InnerVolumeSpecName "kube-api-access-6n5sv". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.504579 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-users" (OuterVolumeSpecName: "sasl-users") pod "fc9beec1-4999-4be9-9dd5-683a7726317e" (UID: "fc9beec1-4999-4be9-9dd5-683a7726317e"). InnerVolumeSpecName "sasl-users". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.505592 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-credentials" (OuterVolumeSpecName: "default-interconnect-inter-router-credentials") pod "fc9beec1-4999-4be9-9dd5-683a7726317e" (UID: "fc9beec1-4999-4be9-9dd5-683a7726317e"). InnerVolumeSpecName "default-interconnect-inter-router-credentials". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.534131 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-m97qq"] Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.534305 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.594766 5109 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-credentials\") on node \"crc\" DevicePath \"\"" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.594806 5109 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-credentials\") on node \"crc\" DevicePath \"\"" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.594815 5109 reconciler_common.go:299] "Volume detached for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-config\") on node \"crc\" DevicePath \"\"" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.594825 5109 reconciler_common.go:299] "Volume detached for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-sasl-users\") on node \"crc\" DevicePath \"\"" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.594834 5109 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-openstack-ca\") on node \"crc\" DevicePath \"\"" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.594846 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-6n5sv\" (UniqueName: \"kubernetes.io/projected/fc9beec1-4999-4be9-9dd5-683a7726317e-kube-api-access-6n5sv\") on node \"crc\" DevicePath \"\"" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.594855 5109 reconciler_common.go:299] "Volume detached for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/fc9beec1-4999-4be9-9dd5-683a7726317e-default-interconnect-inter-router-ca\") on node \"crc\" DevicePath \"\"" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.696068 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/81928491-66a0-4044-8de6-6a44ae591412-sasl-config\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.696139 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-sasl-users\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.696382 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.696455 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.696647 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.696709 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.696795 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjf69\" (UniqueName: \"kubernetes.io/projected/81928491-66a0-4044-8de6-6a44ae591412-kube-api-access-cjf69\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.798068 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-sasl-users\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.798130 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.798156 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.798196 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.798216 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.798232 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-cjf69\" (UniqueName: \"kubernetes.io/projected/81928491-66a0-4044-8de6-6a44ae591412-kube-api-access-cjf69\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.798280 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/81928491-66a0-4044-8de6-6a44ae591412-sasl-config\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.799689 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-config\" (UniqueName: \"kubernetes.io/configmap/81928491-66a0-4044-8de6-6a44ae591412-sasl-config\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.802970 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-credentials\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-openstack-credentials\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.803545 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-openstack-ca\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-openstack-ca\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.803566 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-credentials\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-inter-router-credentials\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.814255 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sasl-users\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-sasl-users\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.817334 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjf69\" (UniqueName: \"kubernetes.io/projected/81928491-66a0-4044-8de6-6a44ae591412-kube-api-access-cjf69\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.827614 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-inter-router-ca\" (UniqueName: \"kubernetes.io/secret/81928491-66a0-4044-8de6-6a44ae591412-default-interconnect-inter-router-ca\") pod \"default-interconnect-55bf8d5cb-m97qq\" (UID: \"81928491-66a0-4044-8de6-6a44ae591412\") " pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.865906 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.968132 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" event={"ID":"fc9beec1-4999-4be9-9dd5-683a7726317e","Type":"ContainerDied","Data":"0f642fc516485d08539e3d0a7cc108e13e0dcc3c52fde47e8509b7b47c7c033f"} Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.968416 5109 scope.go:117] "RemoveContainer" containerID="3484498e803b5859f17664767a8586f5f21a07a2f85b647d75bafb8b3529b1a5" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.968566 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/default-interconnect-55bf8d5cb-dqcvs" Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.997206 5109 generic.go:358] "Generic (PLEG): container finished" podID="d251c432-204d-4df1-a193-fb361eba6102" containerID="b65fce5b231a06dcfc337ca08c0d58faad359207a2dd28028bc4047249326b99" exitCode=0 Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.997327 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerDied","Data":"b65fce5b231a06dcfc337ca08c0d58faad359207a2dd28028bc4047249326b99"} Dec 11 17:10:22 crc kubenswrapper[5109]: I1211 17:10:22.997817 5109 scope.go:117] "RemoveContainer" containerID="b65fce5b231a06dcfc337ca08c0d58faad359207a2dd28028bc4047249326b99" Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.015067 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-dqcvs"] Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.018864 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-dqcvs"] Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.026773 5109 generic.go:358] "Generic (PLEG): container finished" podID="cbadc371-418d-4601-90c4-1d5d3bd787e7" containerID="6e689757652cfc3813f922639c3fdb06df5d0c4b5da6d36f1aa306bcdda9e732" exitCode=0 Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.026913 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" event={"ID":"cbadc371-418d-4601-90c4-1d5d3bd787e7","Type":"ContainerDied","Data":"6e689757652cfc3813f922639c3fdb06df5d0c4b5da6d36f1aa306bcdda9e732"} Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.027183 5109 scope.go:117] "RemoveContainer" containerID="6e689757652cfc3813f922639c3fdb06df5d0c4b5da6d36f1aa306bcdda9e732" Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.054932 5109 generic.go:358] "Generic (PLEG): container finished" podID="80337a91-4532-4bc3-b340-39ba1681f3a1" containerID="3d77ad790743ed319d5abf5bd0bf2a2dc023bc6ba6e624f2e453d8f3808e0ea3" exitCode=0 Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.055101 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerDied","Data":"3d77ad790743ed319d5abf5bd0bf2a2dc023bc6ba6e624f2e453d8f3808e0ea3"} Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.057244 5109 scope.go:117] "RemoveContainer" containerID="3d77ad790743ed319d5abf5bd0bf2a2dc023bc6ba6e624f2e453d8f3808e0ea3" Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.086933 5109 generic.go:358] "Generic (PLEG): container finished" podID="250da305-9832-4e46-9a23-7edcfa9d6682" containerID="9b00465312c11878d4bb79e4c15cbf622f48f2d0fae2c0f6c39d3e1904433be8" exitCode=0 Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.087029 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" event={"ID":"250da305-9832-4e46-9a23-7edcfa9d6682","Type":"ContainerDied","Data":"9b00465312c11878d4bb79e4c15cbf622f48f2d0fae2c0f6c39d3e1904433be8"} Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.087468 5109 scope.go:117] "RemoveContainer" containerID="9b00465312c11878d4bb79e4c15cbf622f48f2d0fae2c0f6c39d3e1904433be8" Dec 11 17:10:23 crc kubenswrapper[5109]: I1211 17:10:23.369613 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/default-interconnect-55bf8d5cb-m97qq"] Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.095337 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerStarted","Data":"6c5957a012c3e283cc503ec1f2baa00bd7bba27701062386fb358add7a32b214"} Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.098100 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerStarted","Data":"4004c3fcdb0c97a9c6981b16898771ef8a6564c7b3f74b57ad9005b4a1776eed"} Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.100340 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" event={"ID":"250da305-9832-4e46-9a23-7edcfa9d6682","Type":"ContainerStarted","Data":"167c0966fd8fc13ef3e89a8f148eebe8d1c3d55dbbfe42a5cbecfb4a9cddea57"} Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.103689 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" event={"ID":"81928491-66a0-4044-8de6-6a44ae591412","Type":"ContainerStarted","Data":"85d03a7d59f5893f74b2eb9a573b3d6bd4b9342a965eee9e89fedc51e19723df"} Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.103731 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" event={"ID":"81928491-66a0-4044-8de6-6a44ae591412","Type":"ContainerStarted","Data":"88e7ea88a0e01dedc5ccfa072e9fff94a98dfc3daaf459ba389f28f25f450cb9"} Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.119323 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" podStartSLOduration=11.427014274 podStartE2EDuration="29.119302253s" podCreationTimestamp="2025-12-11 17:09:55 +0000 UTC" firstStartedPulling="2025-12-11 17:10:05.041852609 +0000 UTC m=+1042.721544075" lastFinishedPulling="2025-12-11 17:10:22.734140588 +0000 UTC m=+1060.413832054" observedRunningTime="2025-12-11 17:10:24.11181398 +0000 UTC m=+1061.791505466" watchObservedRunningTime="2025-12-11 17:10:24.119302253 +0000 UTC m=+1061.798993729" Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.147549 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/default-interconnect-55bf8d5cb-m97qq" podStartSLOduration=3.147531357 podStartE2EDuration="3.147531357s" podCreationTimestamp="2025-12-11 17:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 17:10:24.14439347 +0000 UTC m=+1061.824084956" watchObservedRunningTime="2025-12-11 17:10:24.147531357 +0000 UTC m=+1061.827222823" Dec 11 17:10:24 crc kubenswrapper[5109]: I1211 17:10:24.908866 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc9beec1-4999-4be9-9dd5-683a7726317e" path="/var/lib/kubelet/pods/fc9beec1-4999-4be9-9dd5-683a7726317e/volumes" Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.112223 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerStarted","Data":"d7dc0b133144510e65a262326037029cbc221bd97c614afc091ea36fa58dc653"} Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.117007 5109 generic.go:358] "Generic (PLEG): container finished" podID="e0ad6477-4b84-4aae-9b44-7ca1cbc17690" containerID="6c5957a012c3e283cc503ec1f2baa00bd7bba27701062386fb358add7a32b214" exitCode=0 Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.117047 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerDied","Data":"6c5957a012c3e283cc503ec1f2baa00bd7bba27701062386fb358add7a32b214"} Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.117095 5109 scope.go:117] "RemoveContainer" containerID="4d45043e287b4ba0050304015feca494fc9d187776b5432dff83cd00b0e7ea95" Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.117919 5109 scope.go:117] "RemoveContainer" containerID="6c5957a012c3e283cc503ec1f2baa00bd7bba27701062386fb358add7a32b214" Dec 11 17:10:25 crc kubenswrapper[5109]: E1211 17:10:25.118337 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65_service-telemetry(e0ad6477-4b84-4aae-9b44-7ca1cbc17690)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" podUID="e0ad6477-4b84-4aae-9b44-7ca1cbc17690" Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.122644 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" event={"ID":"cbadc371-418d-4601-90c4-1d5d3bd787e7","Type":"ContainerStarted","Data":"762b045fe047dbcae01eab170a4ea3ec913aeb36d89a1e82b3206f0cdb17f957"} Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.126154 5109 generic.go:358] "Generic (PLEG): container finished" podID="80337a91-4532-4bc3-b340-39ba1681f3a1" containerID="4004c3fcdb0c97a9c6981b16898771ef8a6564c7b3f74b57ad9005b4a1776eed" exitCode=0 Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.126286 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerDied","Data":"4004c3fcdb0c97a9c6981b16898771ef8a6564c7b3f74b57ad9005b4a1776eed"} Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.126753 5109 scope.go:117] "RemoveContainer" containerID="4004c3fcdb0c97a9c6981b16898771ef8a6564c7b3f74b57ad9005b4a1776eed" Dec 11 17:10:25 crc kubenswrapper[5109]: E1211 17:10:25.127013 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56_service-telemetry(80337a91-4532-4bc3-b340-39ba1681f3a1)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" podUID="80337a91-4532-4bc3-b340-39ba1681f3a1" Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.137217 5109 generic.go:358] "Generic (PLEG): container finished" podID="250da305-9832-4e46-9a23-7edcfa9d6682" containerID="167c0966fd8fc13ef3e89a8f148eebe8d1c3d55dbbfe42a5cbecfb4a9cddea57" exitCode=0 Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.137435 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" event={"ID":"250da305-9832-4e46-9a23-7edcfa9d6682","Type":"ContainerDied","Data":"167c0966fd8fc13ef3e89a8f148eebe8d1c3d55dbbfe42a5cbecfb4a9cddea57"} Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.137606 5109 scope.go:117] "RemoveContainer" containerID="167c0966fd8fc13ef3e89a8f148eebe8d1c3d55dbbfe42a5cbecfb4a9cddea57" Dec 11 17:10:25 crc kubenswrapper[5109]: E1211 17:10:25.137970 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc_service-telemetry(250da305-9832-4e46-9a23-7edcfa9d6682)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" podUID="250da305-9832-4e46-9a23-7edcfa9d6682" Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.175987 5109 scope.go:117] "RemoveContainer" containerID="3d77ad790743ed319d5abf5bd0bf2a2dc023bc6ba6e624f2e453d8f3808e0ea3" Dec 11 17:10:25 crc kubenswrapper[5109]: I1211 17:10:25.242685 5109 scope.go:117] "RemoveContainer" containerID="9b00465312c11878d4bb79e4c15cbf622f48f2d0fae2c0f6c39d3e1904433be8" Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.148162 5109 scope.go:117] "RemoveContainer" containerID="167c0966fd8fc13ef3e89a8f148eebe8d1c3d55dbbfe42a5cbecfb4a9cddea57" Dec 11 17:10:26 crc kubenswrapper[5109]: E1211 17:10:26.148686 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc_service-telemetry(250da305-9832-4e46-9a23-7edcfa9d6682)\"" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" podUID="250da305-9832-4e46-9a23-7edcfa9d6682" Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.150698 5109 generic.go:358] "Generic (PLEG): container finished" podID="d251c432-204d-4df1-a193-fb361eba6102" containerID="d7dc0b133144510e65a262326037029cbc221bd97c614afc091ea36fa58dc653" exitCode=0 Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.150896 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerDied","Data":"d7dc0b133144510e65a262326037029cbc221bd97c614afc091ea36fa58dc653"} Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.150947 5109 scope.go:117] "RemoveContainer" containerID="b65fce5b231a06dcfc337ca08c0d58faad359207a2dd28028bc4047249326b99" Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.151313 5109 scope.go:117] "RemoveContainer" containerID="d7dc0b133144510e65a262326037029cbc221bd97c614afc091ea36fa58dc653" Dec 11 17:10:26 crc kubenswrapper[5109]: E1211 17:10:26.151574 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-787645d794-45cc9_service-telemetry(d251c432-204d-4df1-a193-fb361eba6102)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" podUID="d251c432-204d-4df1-a193-fb361eba6102" Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.155120 5109 scope.go:117] "RemoveContainer" containerID="6c5957a012c3e283cc503ec1f2baa00bd7bba27701062386fb358add7a32b214" Dec 11 17:10:26 crc kubenswrapper[5109]: E1211 17:10:26.155546 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65_service-telemetry(e0ad6477-4b84-4aae-9b44-7ca1cbc17690)\"" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" podUID="e0ad6477-4b84-4aae-9b44-7ca1cbc17690" Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.156157 5109 generic.go:358] "Generic (PLEG): container finished" podID="cbadc371-418d-4601-90c4-1d5d3bd787e7" containerID="762b045fe047dbcae01eab170a4ea3ec913aeb36d89a1e82b3206f0cdb17f957" exitCode=0 Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.156241 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" event={"ID":"cbadc371-418d-4601-90c4-1d5d3bd787e7","Type":"ContainerDied","Data":"762b045fe047dbcae01eab170a4ea3ec913aeb36d89a1e82b3206f0cdb17f957"} Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.156472 5109 scope.go:117] "RemoveContainer" containerID="762b045fe047dbcae01eab170a4ea3ec913aeb36d89a1e82b3206f0cdb17f957" Dec 11 17:10:26 crc kubenswrapper[5109]: E1211 17:10:26.156672 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj_service-telemetry(cbadc371-418d-4601-90c4-1d5d3bd787e7)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" podUID="cbadc371-418d-4601-90c4-1d5d3bd787e7" Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.158254 5109 scope.go:117] "RemoveContainer" containerID="4004c3fcdb0c97a9c6981b16898771ef8a6564c7b3f74b57ad9005b4a1776eed" Dec 11 17:10:26 crc kubenswrapper[5109]: E1211 17:10:26.158396 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56_service-telemetry(80337a91-4532-4bc3-b340-39ba1681f3a1)\"" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" podUID="80337a91-4532-4bc3-b340-39ba1681f3a1" Dec 11 17:10:26 crc kubenswrapper[5109]: I1211 17:10:26.278378 5109 scope.go:117] "RemoveContainer" containerID="6e689757652cfc3813f922639c3fdb06df5d0c4b5da6d36f1aa306bcdda9e732" Dec 11 17:10:27 crc kubenswrapper[5109]: I1211 17:10:27.175965 5109 scope.go:117] "RemoveContainer" containerID="d7dc0b133144510e65a262326037029cbc221bd97c614afc091ea36fa58dc653" Dec 11 17:10:27 crc kubenswrapper[5109]: E1211 17:10:27.176187 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-meter-smartgateway-787645d794-45cc9_service-telemetry(d251c432-204d-4df1-a193-fb361eba6102)\"" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" podUID="d251c432-204d-4df1-a193-fb361eba6102" Dec 11 17:10:27 crc kubenswrapper[5109]: I1211 17:10:27.180098 5109 scope.go:117] "RemoveContainer" containerID="762b045fe047dbcae01eab170a4ea3ec913aeb36d89a1e82b3206f0cdb17f957" Dec 11 17:10:27 crc kubenswrapper[5109]: E1211 17:10:27.180318 5109 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"bridge\" with CrashLoopBackOff: \"back-off 10s restarting failed container=bridge pod=default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj_service-telemetry(cbadc371-418d-4601-90c4-1d5d3bd787e7)\"" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" podUID="cbadc371-418d-4601-90c4-1d5d3bd787e7" Dec 11 17:10:37 crc kubenswrapper[5109]: I1211 17:10:37.762468 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:10:37 crc kubenswrapper[5109]: I1211 17:10:37.762811 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:10:37 crc kubenswrapper[5109]: I1211 17:10:37.900843 5109 scope.go:117] "RemoveContainer" containerID="4004c3fcdb0c97a9c6981b16898771ef8a6564c7b3f74b57ad9005b4a1776eed" Dec 11 17:10:37 crc kubenswrapper[5109]: I1211 17:10:37.900989 5109 scope.go:117] "RemoveContainer" containerID="6c5957a012c3e283cc503ec1f2baa00bd7bba27701062386fb358add7a32b214" Dec 11 17:10:37 crc kubenswrapper[5109]: I1211 17:10:37.901302 5109 scope.go:117] "RemoveContainer" containerID="167c0966fd8fc13ef3e89a8f148eebe8d1c3d55dbbfe42a5cbecfb4a9cddea57" Dec 11 17:10:39 crc kubenswrapper[5109]: I1211 17:10:39.900463 5109 scope.go:117] "RemoveContainer" containerID="762b045fe047dbcae01eab170a4ea3ec913aeb36d89a1e82b3206f0cdb17f957" Dec 11 17:10:39 crc kubenswrapper[5109]: I1211 17:10:39.906865 5109 scope.go:117] "RemoveContainer" containerID="d7dc0b133144510e65a262326037029cbc221bd97c614afc091ea36fa58dc653" Dec 11 17:10:41 crc kubenswrapper[5109]: I1211 17:10:41.291884 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-meter-smartgateway-787645d794-45cc9" event={"ID":"d251c432-204d-4df1-a193-fb361eba6102","Type":"ContainerStarted","Data":"a1c06abaf1a8618a25176c235a85726178318c022d3aed7c4096f5ac0937869f"} Dec 11 17:10:41 crc kubenswrapper[5109]: I1211 17:10:41.294685 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-meter-smartgateway-545b564d9f-kvn65" event={"ID":"e0ad6477-4b84-4aae-9b44-7ca1cbc17690","Type":"ContainerStarted","Data":"604c1dbc30a23a60385e1780fa01510950556b8de9de4f239f7f3dabc41e78a9"} Dec 11 17:10:41 crc kubenswrapper[5109]: I1211 17:10:41.296784 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-coll-event-smartgateway-79d7658d89-nzqjj" event={"ID":"cbadc371-418d-4601-90c4-1d5d3bd787e7","Type":"ContainerStarted","Data":"33d3a05c4ac05729252f8efcf9675e9ce210f2954e88b0e1edfb5efc87dce8bb"} Dec 11 17:10:41 crc kubenswrapper[5109]: I1211 17:10:41.299340 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-sens-meter-smartgateway-66d5b7c5fc-lvg56" event={"ID":"80337a91-4532-4bc3-b340-39ba1681f3a1","Type":"ContainerStarted","Data":"8668951c70b22b3b9b142e1575956c86ceb38bc90ef342886af0ae8bcd15bda2"} Dec 11 17:10:41 crc kubenswrapper[5109]: I1211 17:10:41.301618 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/default-cloud1-ceil-event-smartgateway-766ccfb644-b5bxc" event={"ID":"250da305-9832-4e46-9a23-7edcfa9d6682","Type":"ContainerStarted","Data":"dc723233b5f82896f6c8a9c7cf86a013cefb71a712a62720a758501fa051ba8e"} Dec 11 17:10:46 crc kubenswrapper[5109]: I1211 17:10:46.312476 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43820: no serving certificate available for the kubelet" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.190423 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/qdr-test"] Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.196263 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.199850 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"service-telemetry\"/\"default-interconnect-selfsigned\"" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.201439 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"qdr-test-config\"" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.210350 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.314343 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjv62\" (UniqueName: \"kubernetes.io/projected/e698825f-27e1-4099-935f-32ca6fce3b04-kube-api-access-xjv62\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.314485 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/e698825f-27e1-4099-935f-32ca6fce3b04-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.314547 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/e698825f-27e1-4099-935f-32ca6fce3b04-qdr-test-config\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.415324 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/e698825f-27e1-4099-935f-32ca6fce3b04-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.415366 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/e698825f-27e1-4099-935f-32ca6fce3b04-qdr-test-config\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.415425 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-xjv62\" (UniqueName: \"kubernetes.io/projected/e698825f-27e1-4099-935f-32ca6fce3b04-kube-api-access-xjv62\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.416707 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"qdr-test-config\" (UniqueName: \"kubernetes.io/configmap/e698825f-27e1-4099-935f-32ca6fce3b04-qdr-test-config\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.420914 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"default-interconnect-selfsigned-cert\" (UniqueName: \"kubernetes.io/secret/e698825f-27e1-4099-935f-32ca6fce3b04-default-interconnect-selfsigned-cert\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.432496 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjv62\" (UniqueName: \"kubernetes.io/projected/e698825f-27e1-4099-935f-32ca6fce3b04-kube-api-access-xjv62\") pod \"qdr-test\" (UID: \"e698825f-27e1-4099-935f-32ca6fce3b04\") " pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.516222 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/qdr-test" Dec 11 17:10:53 crc kubenswrapper[5109]: I1211 17:10:53.957767 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/qdr-test"] Dec 11 17:10:54 crc kubenswrapper[5109]: I1211 17:10:54.397212 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"e698825f-27e1-4099-935f-32ca6fce3b04","Type":"ContainerStarted","Data":"a886e2de5d31ed5eff226b66a6d77d1043d70bca97a9f319a235f089362a1d89"} Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.442975 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/qdr-test" event={"ID":"e698825f-27e1-4099-935f-32ca6fce3b04","Type":"ContainerStarted","Data":"2faada36234572dd0ccd61642897378c6a5c1efb6fdc496a0189a9c9a7cf36c0"} Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.790461 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/qdr-test" podStartSLOduration=2.083255596 podStartE2EDuration="8.790439112s" podCreationTimestamp="2025-12-11 17:10:53 +0000 UTC" firstStartedPulling="2025-12-11 17:10:53.965962924 +0000 UTC m=+1091.645654420" lastFinishedPulling="2025-12-11 17:11:00.67314645 +0000 UTC m=+1098.352837936" observedRunningTime="2025-12-11 17:11:01.464732997 +0000 UTC m=+1099.144424463" watchObservedRunningTime="2025-12-11 17:11:01.790439112 +0000 UTC m=+1099.470130578" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.795460 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/stf-smoketest-smoke1-829l4"] Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.805877 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-829l4"] Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.806042 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.808045 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-sensubility-config\"" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.810501 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-healthcheck-log\"" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.810659 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-collectd-config\"" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.810926 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-ceilometer-publisher\"" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.811173 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-collectd-entrypoint-script\"" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.811393 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"service-telemetry\"/\"stf-smoketest-ceilometer-entrypoint-script\"" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.942166 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-healthcheck-log\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.942212 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-sensubility-config\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.942258 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.942332 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvxb9\" (UniqueName: \"kubernetes.io/projected/78a7bad5-207d-409b-812d-f101baea9512-kube-api-access-jvxb9\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.942475 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.942582 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-config\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:01 crc kubenswrapper[5109]: I1211 17:11:01.942622 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-publisher\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.044095 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-publisher\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.044337 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-healthcheck-log\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.044975 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-sensubility-config\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.045179 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.045222 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-publisher\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.045278 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-jvxb9\" (UniqueName: \"kubernetes.io/projected/78a7bad5-207d-409b-812d-f101baea9512-kube-api-access-jvxb9\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.045370 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.045530 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-config\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.045608 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-sensubility-config\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.045953 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-healthcheck-log\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.046130 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-entrypoint-script\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.046341 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-config\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.046586 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-entrypoint-script\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.065708 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvxb9\" (UniqueName: \"kubernetes.io/projected/78a7bad5-207d-409b-812d-f101baea9512-kube-api-access-jvxb9\") pod \"stf-smoketest-smoke1-829l4\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.135087 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.311474 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/curl"] Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.320159 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.320356 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.426983 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/stf-smoketest-smoke1-829l4"] Dec 11 17:11:02 crc kubenswrapper[5109]: W1211 17:11:02.433662 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78a7bad5_207d_409b_812d_f101baea9512.slice/crio-879b93d89314d83a70ab175cd0ea7f2527d8561f5adbad630272b254ead12b10 WatchSource:0}: Error finding container 879b93d89314d83a70ab175cd0ea7f2527d8561f5adbad630272b254ead12b10: Status 404 returned error can't find the container with id 879b93d89314d83a70ab175cd0ea7f2527d8561f5adbad630272b254ead12b10 Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.450974 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtcmr\" (UniqueName: \"kubernetes.io/projected/1fbc1401-211b-4cd2-a1b1-bde79a655f1a-kube-api-access-vtcmr\") pod \"curl\" (UID: \"1fbc1401-211b-4cd2-a1b1-bde79a655f1a\") " pod="service-telemetry/curl" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.453538 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-829l4" event={"ID":"78a7bad5-207d-409b-812d-f101baea9512","Type":"ContainerStarted","Data":"879b93d89314d83a70ab175cd0ea7f2527d8561f5adbad630272b254ead12b10"} Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.552501 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-vtcmr\" (UniqueName: \"kubernetes.io/projected/1fbc1401-211b-4cd2-a1b1-bde79a655f1a-kube-api-access-vtcmr\") pod \"curl\" (UID: \"1fbc1401-211b-4cd2-a1b1-bde79a655f1a\") " pod="service-telemetry/curl" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.579933 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtcmr\" (UniqueName: \"kubernetes.io/projected/1fbc1401-211b-4cd2-a1b1-bde79a655f1a-kube-api-access-vtcmr\") pod \"curl\" (UID: \"1fbc1401-211b-4cd2-a1b1-bde79a655f1a\") " pod="service-telemetry/curl" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.656724 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 11 17:11:02 crc kubenswrapper[5109]: I1211 17:11:02.913433 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/curl"] Dec 11 17:11:02 crc kubenswrapper[5109]: W1211 17:11:02.916014 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1fbc1401_211b_4cd2_a1b1_bde79a655f1a.slice/crio-8cb647ea9e9a4f93517c0426a348bf9730ea0061d150ccf78173a74b3e15a991 WatchSource:0}: Error finding container 8cb647ea9e9a4f93517c0426a348bf9730ea0061d150ccf78173a74b3e15a991: Status 404 returned error can't find the container with id 8cb647ea9e9a4f93517c0426a348bf9730ea0061d150ccf78173a74b3e15a991 Dec 11 17:11:03 crc kubenswrapper[5109]: I1211 17:11:03.462414 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"1fbc1401-211b-4cd2-a1b1-bde79a655f1a","Type":"ContainerStarted","Data":"8cb647ea9e9a4f93517c0426a348bf9730ea0061d150ccf78173a74b3e15a991"} Dec 11 17:11:07 crc kubenswrapper[5109]: I1211 17:11:07.761863 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:11:07 crc kubenswrapper[5109]: I1211 17:11:07.762457 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:11:11 crc kubenswrapper[5109]: I1211 17:11:11.519028 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-829l4" event={"ID":"78a7bad5-207d-409b-812d-f101baea9512","Type":"ContainerStarted","Data":"2cfeabadaf7c801e66b64c7400efc6474b56921d6716ae035f6a3ad16a7e5e8c"} Dec 11 17:11:11 crc kubenswrapper[5109]: I1211 17:11:11.521404 5109 generic.go:358] "Generic (PLEG): container finished" podID="1fbc1401-211b-4cd2-a1b1-bde79a655f1a" containerID="62288681447f7d00d1507cd031ac1eb0a67d5870002c37eb097c8b7781e112da" exitCode=0 Dec 11 17:11:11 crc kubenswrapper[5109]: I1211 17:11:11.521523 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"1fbc1401-211b-4cd2-a1b1-bde79a655f1a","Type":"ContainerDied","Data":"62288681447f7d00d1507cd031ac1eb0a67d5870002c37eb097c8b7781e112da"} Dec 11 17:11:12 crc kubenswrapper[5109]: I1211 17:11:12.789239 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 11 17:11:12 crc kubenswrapper[5109]: I1211 17:11:12.925312 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtcmr\" (UniqueName: \"kubernetes.io/projected/1fbc1401-211b-4cd2-a1b1-bde79a655f1a-kube-api-access-vtcmr\") pod \"1fbc1401-211b-4cd2-a1b1-bde79a655f1a\" (UID: \"1fbc1401-211b-4cd2-a1b1-bde79a655f1a\") " Dec 11 17:11:12 crc kubenswrapper[5109]: I1211 17:11:12.930702 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1fbc1401-211b-4cd2-a1b1-bde79a655f1a-kube-api-access-vtcmr" (OuterVolumeSpecName: "kube-api-access-vtcmr") pod "1fbc1401-211b-4cd2-a1b1-bde79a655f1a" (UID: "1fbc1401-211b-4cd2-a1b1-bde79a655f1a"). InnerVolumeSpecName "kube-api-access-vtcmr". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:11:12 crc kubenswrapper[5109]: I1211 17:11:12.982474 5109 ???:1] "http: TLS handshake error from 192.168.126.11:60340: no serving certificate available for the kubelet" Dec 11 17:11:13 crc kubenswrapper[5109]: I1211 17:11:13.026811 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-vtcmr\" (UniqueName: \"kubernetes.io/projected/1fbc1401-211b-4cd2-a1b1-bde79a655f1a-kube-api-access-vtcmr\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:13 crc kubenswrapper[5109]: I1211 17:11:13.250628 5109 ???:1] "http: TLS handshake error from 192.168.126.11:60346: no serving certificate available for the kubelet" Dec 11 17:11:13 crc kubenswrapper[5109]: I1211 17:11:13.536203 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/curl" event={"ID":"1fbc1401-211b-4cd2-a1b1-bde79a655f1a","Type":"ContainerDied","Data":"8cb647ea9e9a4f93517c0426a348bf9730ea0061d150ccf78173a74b3e15a991"} Dec 11 17:11:13 crc kubenswrapper[5109]: I1211 17:11:13.536245 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cb647ea9e9a4f93517c0426a348bf9730ea0061d150ccf78173a74b3e15a991" Dec 11 17:11:13 crc kubenswrapper[5109]: I1211 17:11:13.536312 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/curl" Dec 11 17:11:17 crc kubenswrapper[5109]: I1211 17:11:17.570937 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-829l4" event={"ID":"78a7bad5-207d-409b-812d-f101baea9512","Type":"ContainerStarted","Data":"8c23398a15ae14f259a14193d5f368bbc7374e53c5dff315d9fc7514187ae348"} Dec 11 17:11:17 crc kubenswrapper[5109]: I1211 17:11:17.604432 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/stf-smoketest-smoke1-829l4" podStartSLOduration=2.338059709 podStartE2EDuration="16.604402449s" podCreationTimestamp="2025-12-11 17:11:01 +0000 UTC" firstStartedPulling="2025-12-11 17:11:02.436772668 +0000 UTC m=+1100.116464134" lastFinishedPulling="2025-12-11 17:11:16.703115408 +0000 UTC m=+1114.382806874" observedRunningTime="2025-12-11 17:11:17.601926048 +0000 UTC m=+1115.281617544" watchObservedRunningTime="2025-12-11 17:11:17.604402449 +0000 UTC m=+1115.284093955" Dec 11 17:11:37 crc kubenswrapper[5109]: I1211 17:11:37.761917 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:11:37 crc kubenswrapper[5109]: I1211 17:11:37.762460 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:11:37 crc kubenswrapper[5109]: I1211 17:11:37.762524 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 17:11:37 crc kubenswrapper[5109]: I1211 17:11:37.763419 5109 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4c415588c2932db1c71d652e0d9ee8b8cbc78b73b5bf8bdb07993c476779114d"} pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 17:11:37 crc kubenswrapper[5109]: I1211 17:11:37.763504 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" containerID="cri-o://4c415588c2932db1c71d652e0d9ee8b8cbc78b73b5bf8bdb07993c476779114d" gracePeriod=600 Dec 11 17:11:40 crc kubenswrapper[5109]: I1211 17:11:40.750559 5109 generic.go:358] "Generic (PLEG): container finished" podID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerID="4c415588c2932db1c71d652e0d9ee8b8cbc78b73b5bf8bdb07993c476779114d" exitCode=0 Dec 11 17:11:40 crc kubenswrapper[5109]: I1211 17:11:40.750666 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerDied","Data":"4c415588c2932db1c71d652e0d9ee8b8cbc78b73b5bf8bdb07993c476779114d"} Dec 11 17:11:40 crc kubenswrapper[5109]: I1211 17:11:40.751251 5109 scope.go:117] "RemoveContainer" containerID="1f3b25744dcfda371142af697057ed29ab8b623200bea5beeda0e00d433c8883" Dec 11 17:11:41 crc kubenswrapper[5109]: I1211 17:11:41.759881 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"45536f2d0edfef683866b20ebb97096b8efbbabadf7003013367dc2e49ff2ef9"} Dec 11 17:11:43 crc kubenswrapper[5109]: I1211 17:11:43.407068 5109 ???:1] "http: TLS handshake error from 192.168.126.11:45386: no serving certificate available for the kubelet" Dec 11 17:11:44 crc kubenswrapper[5109]: I1211 17:11:44.802007 5109 generic.go:358] "Generic (PLEG): container finished" podID="78a7bad5-207d-409b-812d-f101baea9512" containerID="2cfeabadaf7c801e66b64c7400efc6474b56921d6716ae035f6a3ad16a7e5e8c" exitCode=0 Dec 11 17:11:44 crc kubenswrapper[5109]: I1211 17:11:44.802066 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-829l4" event={"ID":"78a7bad5-207d-409b-812d-f101baea9512","Type":"ContainerDied","Data":"2cfeabadaf7c801e66b64c7400efc6474b56921d6716ae035f6a3ad16a7e5e8c"} Dec 11 17:11:44 crc kubenswrapper[5109]: I1211 17:11:44.803268 5109 scope.go:117] "RemoveContainer" containerID="2cfeabadaf7c801e66b64c7400efc6474b56921d6716ae035f6a3ad16a7e5e8c" Dec 11 17:11:48 crc kubenswrapper[5109]: I1211 17:11:48.837501 5109 generic.go:358] "Generic (PLEG): container finished" podID="78a7bad5-207d-409b-812d-f101baea9512" containerID="8c23398a15ae14f259a14193d5f368bbc7374e53c5dff315d9fc7514187ae348" exitCode=0 Dec 11 17:11:48 crc kubenswrapper[5109]: I1211 17:11:48.837586 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-829l4" event={"ID":"78a7bad5-207d-409b-812d-f101baea9512","Type":"ContainerDied","Data":"8c23398a15ae14f259a14193d5f368bbc7374e53c5dff315d9fc7514187ae348"} Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.109768 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.657236 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-healthcheck-log\") pod \"78a7bad5-207d-409b-812d-f101baea9512\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.657413 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-sensubility-config\") pod \"78a7bad5-207d-409b-812d-f101baea9512\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.657525 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-entrypoint-script\") pod \"78a7bad5-207d-409b-812d-f101baea9512\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.657622 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-publisher\") pod \"78a7bad5-207d-409b-812d-f101baea9512\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.657682 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvxb9\" (UniqueName: \"kubernetes.io/projected/78a7bad5-207d-409b-812d-f101baea9512-kube-api-access-jvxb9\") pod \"78a7bad5-207d-409b-812d-f101baea9512\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.657791 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-entrypoint-script\") pod \"78a7bad5-207d-409b-812d-f101baea9512\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.657830 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-config\") pod \"78a7bad5-207d-409b-812d-f101baea9512\" (UID: \"78a7bad5-207d-409b-812d-f101baea9512\") " Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.670887 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78a7bad5-207d-409b-812d-f101baea9512-kube-api-access-jvxb9" (OuterVolumeSpecName: "kube-api-access-jvxb9") pod "78a7bad5-207d-409b-812d-f101baea9512" (UID: "78a7bad5-207d-409b-812d-f101baea9512"). InnerVolumeSpecName "kube-api-access-jvxb9". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.685245 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-publisher" (OuterVolumeSpecName: "ceilometer-publisher") pod "78a7bad5-207d-409b-812d-f101baea9512" (UID: "78a7bad5-207d-409b-812d-f101baea9512"). InnerVolumeSpecName "ceilometer-publisher". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.695634 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-entrypoint-script" (OuterVolumeSpecName: "ceilometer-entrypoint-script") pod "78a7bad5-207d-409b-812d-f101baea9512" (UID: "78a7bad5-207d-409b-812d-f101baea9512"). InnerVolumeSpecName "ceilometer-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.698236 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-entrypoint-script" (OuterVolumeSpecName: "collectd-entrypoint-script") pod "78a7bad5-207d-409b-812d-f101baea9512" (UID: "78a7bad5-207d-409b-812d-f101baea9512"). InnerVolumeSpecName "collectd-entrypoint-script". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.701465 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-sensubility-config" (OuterVolumeSpecName: "sensubility-config") pod "78a7bad5-207d-409b-812d-f101baea9512" (UID: "78a7bad5-207d-409b-812d-f101baea9512"). InnerVolumeSpecName "sensubility-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.704409 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-healthcheck-log" (OuterVolumeSpecName: "healthcheck-log") pod "78a7bad5-207d-409b-812d-f101baea9512" (UID: "78a7bad5-207d-409b-812d-f101baea9512"). InnerVolumeSpecName "healthcheck-log". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.705833 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-config" (OuterVolumeSpecName: "collectd-config") pod "78a7bad5-207d-409b-812d-f101baea9512" (UID: "78a7bad5-207d-409b-812d-f101baea9512"). InnerVolumeSpecName "collectd-config". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.759627 5109 reconciler_common.go:299] "Volume detached for volume \"sensubility-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-sensubility-config\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.759672 5109 reconciler_common.go:299] "Volume detached for volume \"ceilometer-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-entrypoint-script\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.759688 5109 reconciler_common.go:299] "Volume detached for volume \"ceilometer-publisher\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-ceilometer-publisher\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.759704 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-jvxb9\" (UniqueName: \"kubernetes.io/projected/78a7bad5-207d-409b-812d-f101baea9512-kube-api-access-jvxb9\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.759716 5109 reconciler_common.go:299] "Volume detached for volume \"collectd-entrypoint-script\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-entrypoint-script\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.759728 5109 reconciler_common.go:299] "Volume detached for volume \"collectd-config\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-collectd-config\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.759758 5109 reconciler_common.go:299] "Volume detached for volume \"healthcheck-log\" (UniqueName: \"kubernetes.io/configmap/78a7bad5-207d-409b-812d-f101baea9512-healthcheck-log\") on node \"crc\" DevicePath \"\"" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.860114 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/stf-smoketest-smoke1-829l4" Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.860612 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/stf-smoketest-smoke1-829l4" event={"ID":"78a7bad5-207d-409b-812d-f101baea9512","Type":"ContainerDied","Data":"879b93d89314d83a70ab175cd0ea7f2527d8561f5adbad630272b254ead12b10"} Dec 11 17:11:50 crc kubenswrapper[5109]: I1211 17:11:50.860682 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="879b93d89314d83a70ab175cd0ea7f2527d8561f5adbad630272b254ead12b10" Dec 11 17:12:08 crc kubenswrapper[5109]: I1211 17:12:08.265517 5109 ???:1] "http: TLS handshake error from 192.168.126.11:39492: no serving certificate available for the kubelet" Dec 11 17:12:13 crc kubenswrapper[5109]: I1211 17:12:13.568408 5109 ???:1] "http: TLS handshake error from 192.168.126.11:39462: no serving certificate available for the kubelet" Dec 11 17:12:43 crc kubenswrapper[5109]: I1211 17:12:43.337625 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 17:12:43 crc kubenswrapper[5109]: I1211 17:12:43.347395 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-authentication_oauth-openshift-d578988d4-qlfht_9607b608-17f2-49a8-a0e3-80610a78b894/oauth-openshift/0.log" Dec 11 17:12:43 crc kubenswrapper[5109]: I1211 17:12:43.373656 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hx2kb_7c45b4ae-02aa-4614-977b-544a9e303bdf/kube-multus/0.log" Dec 11 17:12:43 crc kubenswrapper[5109]: I1211 17:12:43.381818 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-hx2kb_7c45b4ae-02aa-4614-977b-544a9e303bdf/kube-multus/0.log" Dec 11 17:12:43 crc kubenswrapper[5109]: I1211 17:12:43.384781 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 17:12:43 crc kubenswrapper[5109]: I1211 17:12:43.394765 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_9f0bc7fcb0822a2c13eb2d22cd8c0641/kube-controller-manager/0.log" Dec 11 17:12:43 crc kubenswrapper[5109]: I1211 17:12:43.789370 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55638: no serving certificate available for the kubelet" Dec 11 17:13:13 crc kubenswrapper[5109]: I1211 17:13:13.958084 5109 ???:1] "http: TLS handshake error from 192.168.126.11:51430: no serving certificate available for the kubelet" Dec 11 17:13:45 crc kubenswrapper[5109]: I1211 17:13:45.076527 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38182: no serving certificate available for the kubelet" Dec 11 17:13:45 crc kubenswrapper[5109]: I1211 17:13:45.380924 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38186: no serving certificate available for the kubelet" Dec 11 17:13:45 crc kubenswrapper[5109]: I1211 17:13:45.710842 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38196: no serving certificate available for the kubelet" Dec 11 17:13:45 crc kubenswrapper[5109]: I1211 17:13:45.969546 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38198: no serving certificate available for the kubelet" Dec 11 17:13:46 crc kubenswrapper[5109]: I1211 17:13:46.265708 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38202: no serving certificate available for the kubelet" Dec 11 17:13:46 crc kubenswrapper[5109]: I1211 17:13:46.588845 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38214: no serving certificate available for the kubelet" Dec 11 17:13:46 crc kubenswrapper[5109]: I1211 17:13:46.890944 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38222: no serving certificate available for the kubelet" Dec 11 17:13:47 crc kubenswrapper[5109]: I1211 17:13:47.178380 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38232: no serving certificate available for the kubelet" Dec 11 17:13:47 crc kubenswrapper[5109]: I1211 17:13:47.483121 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38238: no serving certificate available for the kubelet" Dec 11 17:13:47 crc kubenswrapper[5109]: I1211 17:13:47.771209 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38246: no serving certificate available for the kubelet" Dec 11 17:13:48 crc kubenswrapper[5109]: I1211 17:13:48.026427 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38260: no serving certificate available for the kubelet" Dec 11 17:13:48 crc kubenswrapper[5109]: I1211 17:13:48.291649 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38262: no serving certificate available for the kubelet" Dec 11 17:13:48 crc kubenswrapper[5109]: I1211 17:13:48.565076 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38266: no serving certificate available for the kubelet" Dec 11 17:13:48 crc kubenswrapper[5109]: I1211 17:13:48.848424 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38268: no serving certificate available for the kubelet" Dec 11 17:13:49 crc kubenswrapper[5109]: I1211 17:13:49.176241 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38284: no serving certificate available for the kubelet" Dec 11 17:13:49 crc kubenswrapper[5109]: I1211 17:13:49.523432 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38288: no serving certificate available for the kubelet" Dec 11 17:13:49 crc kubenswrapper[5109]: I1211 17:13:49.821897 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38300: no serving certificate available for the kubelet" Dec 11 17:13:50 crc kubenswrapper[5109]: I1211 17:13:50.132321 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38306: no serving certificate available for the kubelet" Dec 11 17:13:57 crc kubenswrapper[5109]: I1211 17:13:57.681953 5109 ???:1] "http: TLS handshake error from 192.168.126.11:53802: no serving certificate available for the kubelet" Dec 11 17:14:02 crc kubenswrapper[5109]: I1211 17:14:02.153123 5109 ???:1] "http: TLS handshake error from 192.168.126.11:48274: no serving certificate available for the kubelet" Dec 11 17:14:02 crc kubenswrapper[5109]: I1211 17:14:02.410965 5109 ???:1] "http: TLS handshake error from 192.168.126.11:48290: no serving certificate available for the kubelet" Dec 11 17:14:02 crc kubenswrapper[5109]: I1211 17:14:02.687386 5109 ???:1] "http: TLS handshake error from 192.168.126.11:48300: no serving certificate available for the kubelet" Dec 11 17:14:07 crc kubenswrapper[5109]: I1211 17:14:07.762299 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:14:07 crc kubenswrapper[5109]: I1211 17:14:07.762927 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.389217 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["service-telemetry/infrawatch-operators-q29f7"] Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390498 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="78a7bad5-207d-409b-812d-f101baea9512" containerName="smoketest-collectd" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390512 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a7bad5-207d-409b-812d-f101baea9512" containerName="smoketest-collectd" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390544 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="1fbc1401-211b-4cd2-a1b1-bde79a655f1a" containerName="curl" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390550 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="1fbc1401-211b-4cd2-a1b1-bde79a655f1a" containerName="curl" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390574 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="78a7bad5-207d-409b-812d-f101baea9512" containerName="smoketest-ceilometer" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390581 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="78a7bad5-207d-409b-812d-f101baea9512" containerName="smoketest-ceilometer" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390685 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="78a7bad5-207d-409b-812d-f101baea9512" containerName="smoketest-collectd" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390697 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="78a7bad5-207d-409b-812d-f101baea9512" containerName="smoketest-ceilometer" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.390710 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="1fbc1401-211b-4cd2-a1b1-bde79a655f1a" containerName="curl" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.407068 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-q29f7"] Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.407232 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.504659 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh5gw\" (UniqueName: \"kubernetes.io/projected/844c5b61-9a00-49a5-8006-80b7aad0f309-kube-api-access-mh5gw\") pod \"infrawatch-operators-q29f7\" (UID: \"844c5b61-9a00-49a5-8006-80b7aad0f309\") " pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.605802 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-mh5gw\" (UniqueName: \"kubernetes.io/projected/844c5b61-9a00-49a5-8006-80b7aad0f309-kube-api-access-mh5gw\") pod \"infrawatch-operators-q29f7\" (UID: \"844c5b61-9a00-49a5-8006-80b7aad0f309\") " pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.627607 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh5gw\" (UniqueName: \"kubernetes.io/projected/844c5b61-9a00-49a5-8006-80b7aad0f309-kube-api-access-mh5gw\") pod \"infrawatch-operators-q29f7\" (UID: \"844c5b61-9a00-49a5-8006-80b7aad0f309\") " pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.724185 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.924468 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["service-telemetry/infrawatch-operators-q29f7"] Dec 11 17:14:22 crc kubenswrapper[5109]: I1211 17:14:22.979069 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-q29f7" event={"ID":"844c5b61-9a00-49a5-8006-80b7aad0f309","Type":"ContainerStarted","Data":"44047efede4afa8170c73f620901abbd81d2ce2336da993cc0ae771a30a16cbe"} Dec 11 17:14:23 crc kubenswrapper[5109]: I1211 17:14:23.990448 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-q29f7" event={"ID":"844c5b61-9a00-49a5-8006-80b7aad0f309","Type":"ContainerStarted","Data":"7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30"} Dec 11 17:14:24 crc kubenswrapper[5109]: I1211 17:14:24.005790 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/infrawatch-operators-q29f7" podStartSLOduration=1.264378968 podStartE2EDuration="2.005770091s" podCreationTimestamp="2025-12-11 17:14:22 +0000 UTC" firstStartedPulling="2025-12-11 17:14:22.931043787 +0000 UTC m=+1300.610735253" lastFinishedPulling="2025-12-11 17:14:23.67243491 +0000 UTC m=+1301.352126376" observedRunningTime="2025-12-11 17:14:24.004719226 +0000 UTC m=+1301.684410722" watchObservedRunningTime="2025-12-11 17:14:24.005770091 +0000 UTC m=+1301.685461577" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.309818 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-must-gather-f4psf/must-gather-tv76b"] Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.342404 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f4psf/must-gather-tv76b"] Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.342546 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.344584 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-f4psf\"/\"openshift-service-ca.crt\"" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.345355 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-must-gather-f4psf\"/\"kube-root-ca.crt\"" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.345532 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-must-gather-f4psf\"/\"default-dockercfg-cd4cx\"" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.473756 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db6ae99f-f8a7-45ba-8336-fcd802d3eece-must-gather-output\") pod \"must-gather-tv76b\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.473819 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg4np\" (UniqueName: \"kubernetes.io/projected/db6ae99f-f8a7-45ba-8336-fcd802d3eece-kube-api-access-kg4np\") pod \"must-gather-tv76b\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.574915 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db6ae99f-f8a7-45ba-8336-fcd802d3eece-must-gather-output\") pod \"must-gather-tv76b\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.575237 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-kg4np\" (UniqueName: \"kubernetes.io/projected/db6ae99f-f8a7-45ba-8336-fcd802d3eece-kube-api-access-kg4np\") pod \"must-gather-tv76b\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.575600 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db6ae99f-f8a7-45ba-8336-fcd802d3eece-must-gather-output\") pod \"must-gather-tv76b\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.594113 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg4np\" (UniqueName: \"kubernetes.io/projected/db6ae99f-f8a7-45ba-8336-fcd802d3eece-kube-api-access-kg4np\") pod \"must-gather-tv76b\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.657171 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:14:27 crc kubenswrapper[5109]: I1211 17:14:27.857948 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-f4psf/must-gather-tv76b"] Dec 11 17:14:27 crc kubenswrapper[5109]: W1211 17:14:27.868601 5109 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddb6ae99f_f8a7_45ba_8336_fcd802d3eece.slice/crio-e88e60cc95b2ebff9a30c34e3a6727677f4fdd42530bfd233f96fcab13b68355 WatchSource:0}: Error finding container e88e60cc95b2ebff9a30c34e3a6727677f4fdd42530bfd233f96fcab13b68355: Status 404 returned error can't find the container with id e88e60cc95b2ebff9a30c34e3a6727677f4fdd42530bfd233f96fcab13b68355 Dec 11 17:14:28 crc kubenswrapper[5109]: I1211 17:14:28.020504 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4psf/must-gather-tv76b" event={"ID":"db6ae99f-f8a7-45ba-8336-fcd802d3eece","Type":"ContainerStarted","Data":"e88e60cc95b2ebff9a30c34e3a6727677f4fdd42530bfd233f96fcab13b68355"} Dec 11 17:14:32 crc kubenswrapper[5109]: I1211 17:14:32.724790 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:32 crc kubenswrapper[5109]: I1211 17:14:32.725327 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:32 crc kubenswrapper[5109]: I1211 17:14:32.748168 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:33 crc kubenswrapper[5109]: I1211 17:14:33.087866 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:34 crc kubenswrapper[5109]: I1211 17:14:34.072547 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4psf/must-gather-tv76b" event={"ID":"db6ae99f-f8a7-45ba-8336-fcd802d3eece","Type":"ContainerStarted","Data":"e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb"} Dec 11 17:14:34 crc kubenswrapper[5109]: I1211 17:14:34.072889 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4psf/must-gather-tv76b" event={"ID":"db6ae99f-f8a7-45ba-8336-fcd802d3eece","Type":"ContainerStarted","Data":"353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b"} Dec 11 17:14:34 crc kubenswrapper[5109]: I1211 17:14:34.971917 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-f4psf/must-gather-tv76b" podStartSLOduration=2.6878248559999998 podStartE2EDuration="7.971900599s" podCreationTimestamp="2025-12-11 17:14:27 +0000 UTC" firstStartedPulling="2025-12-11 17:14:27.872610369 +0000 UTC m=+1305.552301825" lastFinishedPulling="2025-12-11 17:14:33.156686102 +0000 UTC m=+1310.836377568" observedRunningTime="2025-12-11 17:14:34.091174694 +0000 UTC m=+1311.770866160" watchObservedRunningTime="2025-12-11 17:14:34.971900599 +0000 UTC m=+1312.651592065" Dec 11 17:14:34 crc kubenswrapper[5109]: I1211 17:14:34.978347 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-q29f7"] Dec 11 17:14:35 crc kubenswrapper[5109]: I1211 17:14:35.078446 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="service-telemetry/infrawatch-operators-q29f7" podUID="844c5b61-9a00-49a5-8006-80b7aad0f309" containerName="registry-server" containerID="cri-o://7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30" gracePeriod=2 Dec 11 17:14:35 crc kubenswrapper[5109]: I1211 17:14:35.422981 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:35 crc kubenswrapper[5109]: I1211 17:14:35.509252 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh5gw\" (UniqueName: \"kubernetes.io/projected/844c5b61-9a00-49a5-8006-80b7aad0f309-kube-api-access-mh5gw\") pod \"844c5b61-9a00-49a5-8006-80b7aad0f309\" (UID: \"844c5b61-9a00-49a5-8006-80b7aad0f309\") " Dec 11 17:14:35 crc kubenswrapper[5109]: I1211 17:14:35.514937 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/844c5b61-9a00-49a5-8006-80b7aad0f309-kube-api-access-mh5gw" (OuterVolumeSpecName: "kube-api-access-mh5gw") pod "844c5b61-9a00-49a5-8006-80b7aad0f309" (UID: "844c5b61-9a00-49a5-8006-80b7aad0f309"). InnerVolumeSpecName "kube-api-access-mh5gw". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:14:35 crc kubenswrapper[5109]: I1211 17:14:35.610168 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-mh5gw\" (UniqueName: \"kubernetes.io/projected/844c5b61-9a00-49a5-8006-80b7aad0f309-kube-api-access-mh5gw\") on node \"crc\" DevicePath \"\"" Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.087944 5109 generic.go:358] "Generic (PLEG): container finished" podID="844c5b61-9a00-49a5-8006-80b7aad0f309" containerID="7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30" exitCode=0 Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.088033 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/infrawatch-operators-q29f7" Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.088059 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-q29f7" event={"ID":"844c5b61-9a00-49a5-8006-80b7aad0f309","Type":"ContainerDied","Data":"7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30"} Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.088477 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="service-telemetry/infrawatch-operators-q29f7" event={"ID":"844c5b61-9a00-49a5-8006-80b7aad0f309","Type":"ContainerDied","Data":"44047efede4afa8170c73f620901abbd81d2ce2336da993cc0ae771a30a16cbe"} Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.088518 5109 scope.go:117] "RemoveContainer" containerID="7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30" Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.113322 5109 scope.go:117] "RemoveContainer" containerID="7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30" Dec 11 17:14:36 crc kubenswrapper[5109]: E1211 17:14:36.115635 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30\": container with ID starting with 7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30 not found: ID does not exist" containerID="7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30" Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.115673 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30"} err="failed to get container status \"7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30\": rpc error: code = NotFound desc = could not find container \"7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30\": container with ID starting with 7eb0ea11ef3b1a83c63827901baec26506c4a7be331ef060b6460e18aba5ab30 not found: ID does not exist" Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.120216 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["service-telemetry/infrawatch-operators-q29f7"] Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.126164 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["service-telemetry/infrawatch-operators-q29f7"] Dec 11 17:14:36 crc kubenswrapper[5109]: I1211 17:14:36.909339 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="844c5b61-9a00-49a5-8006-80b7aad0f309" path="/var/lib/kubelet/pods/844c5b61-9a00-49a5-8006-80b7aad0f309/volumes" Dec 11 17:14:37 crc kubenswrapper[5109]: I1211 17:14:37.305243 5109 ???:1] "http: TLS handshake error from 192.168.126.11:38328: no serving certificate available for the kubelet" Dec 11 17:14:37 crc kubenswrapper[5109]: I1211 17:14:37.762268 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:14:37 crc kubenswrapper[5109]: I1211 17:14:37.762511 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:14:52 crc kubenswrapper[5109]: I1211 17:14:52.129448 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55546: no serving certificate available for the kubelet" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.143913 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q"] Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.145370 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="844c5b61-9a00-49a5-8006-80b7aad0f309" containerName="registry-server" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.145389 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="844c5b61-9a00-49a5-8006-80b7aad0f309" containerName="registry-server" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.145579 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="844c5b61-9a00-49a5-8006-80b7aad0f309" containerName="registry-server" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.159457 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q"] Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.159687 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.162901 5109 reflector.go:430] "Caches populated" type="*v1.Secret" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-dockercfg-vfqp6\"" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.163208 5109 reflector.go:430] "Caches populated" type="*v1.ConfigMap" reflector="object-\"openshift-operator-lifecycle-manager\"/\"collect-profiles-config\"" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.173073 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkbdh\" (UniqueName: \"kubernetes.io/projected/c01044bf-a35c-4f04-898e-61c43e2ad535-kube-api-access-hkbdh\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.173126 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01044bf-a35c-4f04-898e-61c43e2ad535-config-volume\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.173182 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01044bf-a35c-4f04-898e-61c43e2ad535-secret-volume\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.274409 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-hkbdh\" (UniqueName: \"kubernetes.io/projected/c01044bf-a35c-4f04-898e-61c43e2ad535-kube-api-access-hkbdh\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.274466 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01044bf-a35c-4f04-898e-61c43e2ad535-config-volume\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.274495 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01044bf-a35c-4f04-898e-61c43e2ad535-secret-volume\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.276589 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01044bf-a35c-4f04-898e-61c43e2ad535-config-volume\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.281507 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01044bf-a35c-4f04-898e-61c43e2ad535-secret-volume\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.305430 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkbdh\" (UniqueName: \"kubernetes.io/projected/c01044bf-a35c-4f04-898e-61c43e2ad535-kube-api-access-hkbdh\") pod \"collect-profiles-29424555-pr88q\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.488941 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.931267 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q"] Dec 11 17:15:00 crc kubenswrapper[5109]: I1211 17:15:00.938592 5109 provider.go:93] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 11 17:15:01 crc kubenswrapper[5109]: I1211 17:15:01.270805 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" event={"ID":"c01044bf-a35c-4f04-898e-61c43e2ad535","Type":"ContainerStarted","Data":"cabeb2cd01e2631c8e21203d78f43b8ac619a79b8992344cf234f06bb7b8235c"} Dec 11 17:15:01 crc kubenswrapper[5109]: I1211 17:15:01.270874 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" event={"ID":"c01044bf-a35c-4f04-898e-61c43e2ad535","Type":"ContainerStarted","Data":"31f9c4b9672c4878e26bf1cb86381eaf01c69b0badf11547f5863bec0800e0df"} Dec 11 17:15:01 crc kubenswrapper[5109]: I1211 17:15:01.291910 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" podStartSLOduration=1.29188816 podStartE2EDuration="1.29188816s" podCreationTimestamp="2025-12-11 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-11 17:15:01.287374071 +0000 UTC m=+1338.967065537" watchObservedRunningTime="2025-12-11 17:15:01.29188816 +0000 UTC m=+1338.971579636" Dec 11 17:15:02 crc kubenswrapper[5109]: I1211 17:15:02.279548 5109 generic.go:358] "Generic (PLEG): container finished" podID="c01044bf-a35c-4f04-898e-61c43e2ad535" containerID="cabeb2cd01e2631c8e21203d78f43b8ac619a79b8992344cf234f06bb7b8235c" exitCode=0 Dec 11 17:15:02 crc kubenswrapper[5109]: I1211 17:15:02.279655 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" event={"ID":"c01044bf-a35c-4f04-898e-61c43e2ad535","Type":"ContainerDied","Data":"cabeb2cd01e2631c8e21203d78f43b8ac619a79b8992344cf234f06bb7b8235c"} Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.556947 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.620449 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01044bf-a35c-4f04-898e-61c43e2ad535-config-volume\") pod \"c01044bf-a35c-4f04-898e-61c43e2ad535\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.620553 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkbdh\" (UniqueName: \"kubernetes.io/projected/c01044bf-a35c-4f04-898e-61c43e2ad535-kube-api-access-hkbdh\") pod \"c01044bf-a35c-4f04-898e-61c43e2ad535\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.620674 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01044bf-a35c-4f04-898e-61c43e2ad535-secret-volume\") pod \"c01044bf-a35c-4f04-898e-61c43e2ad535\" (UID: \"c01044bf-a35c-4f04-898e-61c43e2ad535\") " Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.621206 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c01044bf-a35c-4f04-898e-61c43e2ad535-config-volume" (OuterVolumeSpecName: "config-volume") pod "c01044bf-a35c-4f04-898e-61c43e2ad535" (UID: "c01044bf-a35c-4f04-898e-61c43e2ad535"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGIDValue "" Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.630001 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c01044bf-a35c-4f04-898e-61c43e2ad535-kube-api-access-hkbdh" (OuterVolumeSpecName: "kube-api-access-hkbdh") pod "c01044bf-a35c-4f04-898e-61c43e2ad535" (UID: "c01044bf-a35c-4f04-898e-61c43e2ad535"). InnerVolumeSpecName "kube-api-access-hkbdh". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.630014 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c01044bf-a35c-4f04-898e-61c43e2ad535-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c01044bf-a35c-4f04-898e-61c43e2ad535" (UID: "c01044bf-a35c-4f04-898e-61c43e2ad535"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGIDValue "" Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.722036 5109 reconciler_common.go:299] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c01044bf-a35c-4f04-898e-61c43e2ad535-config-volume\") on node \"crc\" DevicePath \"\"" Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.722067 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-hkbdh\" (UniqueName: \"kubernetes.io/projected/c01044bf-a35c-4f04-898e-61c43e2ad535-kube-api-access-hkbdh\") on node \"crc\" DevicePath \"\"" Dec 11 17:15:03 crc kubenswrapper[5109]: I1211 17:15:03.722078 5109 reconciler_common.go:299] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c01044bf-a35c-4f04-898e-61c43e2ad535-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 11 17:15:04 crc kubenswrapper[5109]: I1211 17:15:04.293953 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" Dec 11 17:15:04 crc kubenswrapper[5109]: I1211 17:15:04.294008 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29424555-pr88q" event={"ID":"c01044bf-a35c-4f04-898e-61c43e2ad535","Type":"ContainerDied","Data":"31f9c4b9672c4878e26bf1cb86381eaf01c69b0badf11547f5863bec0800e0df"} Dec 11 17:15:04 crc kubenswrapper[5109]: I1211 17:15:04.294074 5109 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="31f9c4b9672c4878e26bf1cb86381eaf01c69b0badf11547f5863bec0800e0df" Dec 11 17:15:07 crc kubenswrapper[5109]: I1211 17:15:07.761538 5109 patch_prober.go:28] interesting pod/machine-config-daemon-xgkvv container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 11 17:15:07 crc kubenswrapper[5109]: I1211 17:15:07.762752 5109 prober.go:120] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 11 17:15:07 crc kubenswrapper[5109]: I1211 17:15:07.762812 5109 kubelet.go:2658] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" Dec 11 17:15:07 crc kubenswrapper[5109]: I1211 17:15:07.763423 5109 kuberuntime_manager.go:1107] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"45536f2d0edfef683866b20ebb97096b8efbbabadf7003013367dc2e49ff2ef9"} pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 11 17:15:07 crc kubenswrapper[5109]: I1211 17:15:07.763477 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" podUID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerName="machine-config-daemon" containerID="cri-o://45536f2d0edfef683866b20ebb97096b8efbbabadf7003013367dc2e49ff2ef9" gracePeriod=600 Dec 11 17:15:08 crc kubenswrapper[5109]: I1211 17:15:08.322239 5109 generic.go:358] "Generic (PLEG): container finished" podID="c1f09ed1-77b5-404d-abd1-44afa5fab784" containerID="45536f2d0edfef683866b20ebb97096b8efbbabadf7003013367dc2e49ff2ef9" exitCode=0 Dec 11 17:15:08 crc kubenswrapper[5109]: I1211 17:15:08.322318 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerDied","Data":"45536f2d0edfef683866b20ebb97096b8efbbabadf7003013367dc2e49ff2ef9"} Dec 11 17:15:08 crc kubenswrapper[5109]: I1211 17:15:08.322598 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-xgkvv" event={"ID":"c1f09ed1-77b5-404d-abd1-44afa5fab784","Type":"ContainerStarted","Data":"a6c082fb3a141449cc735ec7628aac348c537a07f47dcfe508163562c62bb585"} Dec 11 17:15:08 crc kubenswrapper[5109]: I1211 17:15:08.322624 5109 scope.go:117] "RemoveContainer" containerID="4c415588c2932db1c71d652e0d9ee8b8cbc78b73b5bf8bdb07993c476779114d" Dec 11 17:15:10 crc kubenswrapper[5109]: I1211 17:15:10.021322 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55964: no serving certificate available for the kubelet" Dec 11 17:15:10 crc kubenswrapper[5109]: I1211 17:15:10.156849 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55978: no serving certificate available for the kubelet" Dec 11 17:15:10 crc kubenswrapper[5109]: I1211 17:15:10.196422 5109 ???:1] "http: TLS handshake error from 192.168.126.11:55980: no serving certificate available for the kubelet" Dec 11 17:15:21 crc kubenswrapper[5109]: I1211 17:15:21.394076 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35388: no serving certificate available for the kubelet" Dec 11 17:15:21 crc kubenswrapper[5109]: I1211 17:15:21.540120 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35400: no serving certificate available for the kubelet" Dec 11 17:15:21 crc kubenswrapper[5109]: I1211 17:15:21.587320 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35406: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.160359 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43368: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.294518 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43380: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.305997 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43396: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.313389 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43404: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.485825 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43412: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.486297 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43418: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.492419 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43430: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.649363 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43442: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.799336 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43448: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.806103 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43458: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.814518 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43472: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.979298 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43474: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.982825 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43478: no serving certificate available for the kubelet" Dec 11 17:15:35 crc kubenswrapper[5109]: I1211 17:15:35.991896 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43494: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.137865 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43510: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.393152 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43524: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.396680 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43530: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.429734 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43532: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.644571 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43548: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.648869 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43552: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.704772 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43554: no serving certificate available for the kubelet" Dec 11 17:15:36 crc kubenswrapper[5109]: I1211 17:15:36.894376 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43556: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.044411 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43562: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.058150 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43568: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.061504 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43578: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.244292 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43586: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.250312 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43592: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.271530 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43594: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.415541 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43606: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.569226 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43612: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.576104 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43616: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.628190 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43632: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.757683 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43636: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.781092 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43640: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.793031 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43656: no serving certificate available for the kubelet" Dec 11 17:15:37 crc kubenswrapper[5109]: I1211 17:15:37.898645 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43672: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.021588 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43686: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.028860 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43692: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.029260 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43694: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.167320 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43702: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.191465 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43706: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.205578 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43734: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.205883 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43722: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.340360 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43736: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.492249 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43752: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.508577 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43754: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.520401 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43758: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.645036 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43766: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.648935 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43768: no serving certificate available for the kubelet" Dec 11 17:15:38 crc kubenswrapper[5109]: I1211 17:15:38.673113 5109 ???:1] "http: TLS handshake error from 192.168.126.11:43772: no serving certificate available for the kubelet" Dec 11 17:15:49 crc kubenswrapper[5109]: I1211 17:15:49.141698 5109 ???:1] "http: TLS handshake error from 192.168.126.11:36550: no serving certificate available for the kubelet" Dec 11 17:15:49 crc kubenswrapper[5109]: I1211 17:15:49.345065 5109 ???:1] "http: TLS handshake error from 192.168.126.11:36564: no serving certificate available for the kubelet" Dec 11 17:15:49 crc kubenswrapper[5109]: I1211 17:15:49.397236 5109 ???:1] "http: TLS handshake error from 192.168.126.11:36572: no serving certificate available for the kubelet" Dec 11 17:15:49 crc kubenswrapper[5109]: I1211 17:15:49.537194 5109 ???:1] "http: TLS handshake error from 192.168.126.11:36578: no serving certificate available for the kubelet" Dec 11 17:15:49 crc kubenswrapper[5109]: I1211 17:15:49.600962 5109 ???:1] "http: TLS handshake error from 192.168.126.11:36592: no serving certificate available for the kubelet" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.278060 5109 kubelet.go:2537] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vw9qj"] Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.279435 5109 cpu_manager.go:401] "RemoveStaleState: containerMap: removing container" podUID="c01044bf-a35c-4f04-898e-61c43e2ad535" containerName="collect-profiles" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.279455 5109 state_mem.go:107] "Deleted CPUSet assignment" podUID="c01044bf-a35c-4f04-898e-61c43e2ad535" containerName="collect-profiles" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.279602 5109 memory_manager.go:356] "RemoveStaleState removing state" podUID="c01044bf-a35c-4f04-898e-61c43e2ad535" containerName="collect-profiles" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.735938 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vw9qj"] Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.736147 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.872658 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae23a918-1dac-4bf7-9dd7-6b4628650a52-catalog-content\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.873032 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae23a918-1dac-4bf7-9dd7-6b4628650a52-utilities\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.873142 5109 reconciler_common.go:251] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wlk6\" (UniqueName: \"kubernetes.io/projected/ae23a918-1dac-4bf7-9dd7-6b4628650a52-kube-api-access-4wlk6\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.975035 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae23a918-1dac-4bf7-9dd7-6b4628650a52-utilities\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.975114 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"kube-api-access-4wlk6\" (UniqueName: \"kubernetes.io/projected/ae23a918-1dac-4bf7-9dd7-6b4628650a52-kube-api-access-4wlk6\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.975161 5109 reconciler_common.go:224] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae23a918-1dac-4bf7-9dd7-6b4628650a52-catalog-content\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.975639 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ae23a918-1dac-4bf7-9dd7-6b4628650a52-utilities\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:54 crc kubenswrapper[5109]: I1211 17:15:54.975700 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ae23a918-1dac-4bf7-9dd7-6b4628650a52-catalog-content\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:55 crc kubenswrapper[5109]: I1211 17:15:55.000881 5109 operation_generator.go:615] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wlk6\" (UniqueName: \"kubernetes.io/projected/ae23a918-1dac-4bf7-9dd7-6b4628650a52-kube-api-access-4wlk6\") pod \"redhat-operators-vw9qj\" (UID: \"ae23a918-1dac-4bf7-9dd7-6b4628650a52\") " pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:55 crc kubenswrapper[5109]: I1211 17:15:55.053694 5109 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:15:55 crc kubenswrapper[5109]: I1211 17:15:55.533621 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vw9qj"] Dec 11 17:15:55 crc kubenswrapper[5109]: I1211 17:15:55.627005 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw9qj" event={"ID":"ae23a918-1dac-4bf7-9dd7-6b4628650a52","Type":"ContainerStarted","Data":"32daaa747a6c9a556245c4fab3f56125af028c8176f7c0e0f8b44ca5e502c287"} Dec 11 17:15:56 crc kubenswrapper[5109]: I1211 17:15:56.635792 5109 generic.go:358] "Generic (PLEG): container finished" podID="ae23a918-1dac-4bf7-9dd7-6b4628650a52" containerID="b8ed1973c4f82505a5ffbd20db3bf34a276e3478a0b9155f75233880121fae5a" exitCode=0 Dec 11 17:15:56 crc kubenswrapper[5109]: I1211 17:15:56.635897 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw9qj" event={"ID":"ae23a918-1dac-4bf7-9dd7-6b4628650a52","Type":"ContainerDied","Data":"b8ed1973c4f82505a5ffbd20db3bf34a276e3478a0b9155f75233880121fae5a"} Dec 11 17:16:04 crc kubenswrapper[5109]: I1211 17:16:04.692730 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw9qj" event={"ID":"ae23a918-1dac-4bf7-9dd7-6b4628650a52","Type":"ContainerStarted","Data":"d1b7fdb0cf3f3020086aeff8ac815b593eda2705b93786d771abc3aaf32def5d"} Dec 11 17:16:05 crc kubenswrapper[5109]: I1211 17:16:05.700258 5109 generic.go:358] "Generic (PLEG): container finished" podID="ae23a918-1dac-4bf7-9dd7-6b4628650a52" containerID="d1b7fdb0cf3f3020086aeff8ac815b593eda2705b93786d771abc3aaf32def5d" exitCode=0 Dec 11 17:16:05 crc kubenswrapper[5109]: I1211 17:16:05.700320 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw9qj" event={"ID":"ae23a918-1dac-4bf7-9dd7-6b4628650a52","Type":"ContainerDied","Data":"d1b7fdb0cf3f3020086aeff8ac815b593eda2705b93786d771abc3aaf32def5d"} Dec 11 17:16:06 crc kubenswrapper[5109]: I1211 17:16:06.715837 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vw9qj" event={"ID":"ae23a918-1dac-4bf7-9dd7-6b4628650a52","Type":"ContainerStarted","Data":"8f27a59ba1156464e0509d6a78801114f7cf31e43d83559b1f86e809c3a6bab6"} Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.055106 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="not ready" pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.055397 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.102106 5109 kubelet.go:2658] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.119919 5109 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vw9qj" podStartSLOduration=13.565464124 podStartE2EDuration="21.119899974s" podCreationTimestamp="2025-12-11 17:15:54 +0000 UTC" firstStartedPulling="2025-12-11 17:15:56.637251693 +0000 UTC m=+1394.316943149" lastFinishedPulling="2025-12-11 17:16:04.191687533 +0000 UTC m=+1401.871378999" observedRunningTime="2025-12-11 17:16:06.740209532 +0000 UTC m=+1404.419900998" watchObservedRunningTime="2025-12-11 17:16:15.119899974 +0000 UTC m=+1412.799591440" Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.830037 5109 kubelet.go:2658] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vw9qj" Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.905984 5109 kubelet.go:2544] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vw9qj"] Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.942901 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tg6k2"] Dec 11 17:16:15 crc kubenswrapper[5109]: I1211 17:16:15.943180 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-tg6k2" podUID="0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" containerName="registry-server" containerID="cri-o://0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9" gracePeriod=2 Dec 11 17:16:17 crc kubenswrapper[5109]: E1211 17:16:17.218984 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9 is running failed: container process not found" containerID="0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 17:16:17 crc kubenswrapper[5109]: E1211 17:16:17.219769 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9 is running failed: container process not found" containerID="0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 17:16:17 crc kubenswrapper[5109]: E1211 17:16:17.220078 5109 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9 is running failed: container process not found" containerID="0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9" cmd=["grpc_health_probe","-addr=:50051"] Dec 11 17:16:17 crc kubenswrapper[5109]: E1211 17:16:17.220134 5109 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-tg6k2" podUID="0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" containerName="registry-server" probeResult="unknown" Dec 11 17:16:19 crc kubenswrapper[5109]: I1211 17:16:19.818005 5109 generic.go:358] "Generic (PLEG): container finished" podID="0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" containerID="0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9" exitCode=0 Dec 11 17:16:19 crc kubenswrapper[5109]: I1211 17:16:19.818075 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tg6k2" event={"ID":"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6","Type":"ContainerDied","Data":"0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9"} Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.634406 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.762877 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-utilities\") pod \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.762924 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-catalog-content\") pod \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.763038 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rk2bt\" (UniqueName: \"kubernetes.io/projected/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-kube-api-access-rk2bt\") pod \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\" (UID: \"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6\") " Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.764231 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-utilities" (OuterVolumeSpecName: "utilities") pod "0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" (UID: "0ec445aa-ae2d-4903-9a2a-a2e811f43ad6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.783542 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-kube-api-access-rk2bt" (OuterVolumeSpecName: "kube-api-access-rk2bt") pod "0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" (UID: "0ec445aa-ae2d-4903-9a2a-a2e811f43ad6"). InnerVolumeSpecName "kube-api-access-rk2bt". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.835380 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-tg6k2" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.835511 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-tg6k2" event={"ID":"0ec445aa-ae2d-4903-9a2a-a2e811f43ad6","Type":"ContainerDied","Data":"baf12f7fb5b7588928ac4ed5e3d1f480647f07337981e471c16e01c5271bfbac"} Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.835611 5109 scope.go:117] "RemoveContainer" containerID="0778e283deff48bb5a52df0e7aa12e9aa9c5ec3652b9b9d06e5703c4e3d20cb9" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.853600 5109 scope.go:117] "RemoveContainer" containerID="b027c5c536838006934e7164a78b70425652a1dad8e5d5129ec8f0627af5b463" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.855533 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" (UID: "0ec445aa-ae2d-4903-9a2a-a2e811f43ad6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.867696 5109 reconciler_common.go:299] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-utilities\") on node \"crc\" DevicePath \"\"" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.867756 5109 reconciler_common.go:299] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.867772 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-rk2bt\" (UniqueName: \"kubernetes.io/projected/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6-kube-api-access-rk2bt\") on node \"crc\" DevicePath \"\"" Dec 11 17:16:20 crc kubenswrapper[5109]: I1211 17:16:20.875859 5109 scope.go:117] "RemoveContainer" containerID="de9500dea283e6a2d4a5098b6c5e933e663c2369ed3782562923899943d32580" Dec 11 17:16:21 crc kubenswrapper[5109]: I1211 17:16:21.154856 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-tg6k2"] Dec 11 17:16:21 crc kubenswrapper[5109]: I1211 17:16:21.161060 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-tg6k2"] Dec 11 17:16:22 crc kubenswrapper[5109]: I1211 17:16:22.909997 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ec445aa-ae2d-4903-9a2a-a2e811f43ad6" path="/var/lib/kubelet/pods/0ec445aa-ae2d-4903-9a2a-a2e811f43ad6/volumes" Dec 11 17:16:27 crc kubenswrapper[5109]: I1211 17:16:27.890954 5109 generic.go:358] "Generic (PLEG): container finished" podID="db6ae99f-f8a7-45ba-8336-fcd802d3eece" containerID="353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b" exitCode=0 Dec 11 17:16:27 crc kubenswrapper[5109]: I1211 17:16:27.891036 5109 kubelet.go:2569] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-f4psf/must-gather-tv76b" event={"ID":"db6ae99f-f8a7-45ba-8336-fcd802d3eece","Type":"ContainerDied","Data":"353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b"} Dec 11 17:16:27 crc kubenswrapper[5109]: I1211 17:16:27.895001 5109 scope.go:117] "RemoveContainer" containerID="353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.323361 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35328: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.479827 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35330: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.493225 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35338: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.520565 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35352: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.532124 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35360: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.548207 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35368: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.559188 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35376: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.574322 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35382: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.587006 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35388: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.733216 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35400: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.745878 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35402: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.768160 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35410: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.777725 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35412: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.789943 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35422: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.802719 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35436: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.815801 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35452: no serving certificate available for the kubelet" Dec 11 17:16:37 crc kubenswrapper[5109]: I1211 17:16:37.826881 5109 ???:1] "http: TLS handshake error from 192.168.126.11:35458: no serving certificate available for the kubelet" Dec 11 17:16:42 crc kubenswrapper[5109]: I1211 17:16:42.881555 5109 kubelet.go:2553] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-f4psf/must-gather-tv76b"] Dec 11 17:16:42 crc kubenswrapper[5109]: I1211 17:16:42.883716 5109 kuberuntime_container.go:858] "Killing container with a grace period" pod="openshift-must-gather-f4psf/must-gather-tv76b" podUID="db6ae99f-f8a7-45ba-8336-fcd802d3eece" containerName="copy" containerID="cri-o://e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb" gracePeriod=2 Dec 11 17:16:42 crc kubenswrapper[5109]: I1211 17:16:42.885821 5109 status_manager.go:895] "Failed to get status for pod" podUID="db6ae99f-f8a7-45ba-8336-fcd802d3eece" pod="openshift-must-gather-f4psf/must-gather-tv76b" err="pods \"must-gather-tv76b\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-f4psf\": no relationship found between node 'crc' and this object" Dec 11 17:16:42 crc kubenswrapper[5109]: I1211 17:16:42.889505 5109 kubelet.go:2547] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-f4psf/must-gather-tv76b"] Dec 11 17:16:42 crc kubenswrapper[5109]: I1211 17:16:42.911617 5109 status_manager.go:895] "Failed to get status for pod" podUID="db6ae99f-f8a7-45ba-8336-fcd802d3eece" pod="openshift-must-gather-f4psf/must-gather-tv76b" err="pods \"must-gather-tv76b\" is forbidden: User \"system:node:crc\" cannot get resource \"pods\" in API group \"\" in the namespace \"openshift-must-gather-f4psf\": no relationship found between node 'crc' and this object" Dec 11 17:16:43 crc kubenswrapper[5109]: I1211 17:16:43.766954 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f4psf_must-gather-tv76b_db6ae99f-f8a7-45ba-8336-fcd802d3eece/copy/0.log" Dec 11 17:16:43 crc kubenswrapper[5109]: I1211 17:16:43.767910 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:16:43 crc kubenswrapper[5109]: I1211 17:16:43.925933 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db6ae99f-f8a7-45ba-8336-fcd802d3eece-must-gather-output\") pod \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " Dec 11 17:16:43 crc kubenswrapper[5109]: I1211 17:16:43.926282 5109 reconciler_common.go:162] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kg4np\" (UniqueName: \"kubernetes.io/projected/db6ae99f-f8a7-45ba-8336-fcd802d3eece-kube-api-access-kg4np\") pod \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\" (UID: \"db6ae99f-f8a7-45ba-8336-fcd802d3eece\") " Dec 11 17:16:43 crc kubenswrapper[5109]: I1211 17:16:43.934092 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db6ae99f-f8a7-45ba-8336-fcd802d3eece-kube-api-access-kg4np" (OuterVolumeSpecName: "kube-api-access-kg4np") pod "db6ae99f-f8a7-45ba-8336-fcd802d3eece" (UID: "db6ae99f-f8a7-45ba-8336-fcd802d3eece"). InnerVolumeSpecName "kube-api-access-kg4np". PluginName "kubernetes.io/projected", VolumeGIDValue "" Dec 11 17:16:43 crc kubenswrapper[5109]: I1211 17:16:43.966603 5109 operation_generator.go:781] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/db6ae99f-f8a7-45ba-8336-fcd802d3eece-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "db6ae99f-f8a7-45ba-8336-fcd802d3eece" (UID: "db6ae99f-f8a7-45ba-8336-fcd802d3eece"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGIDValue "" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.010100 5109 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-f4psf_must-gather-tv76b_db6ae99f-f8a7-45ba-8336-fcd802d3eece/copy/0.log" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.011087 5109 generic.go:358] "Generic (PLEG): container finished" podID="db6ae99f-f8a7-45ba-8336-fcd802d3eece" containerID="e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb" exitCode=143 Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.011152 5109 scope.go:117] "RemoveContainer" containerID="e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.011163 5109 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-f4psf/must-gather-tv76b" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.028728 5109 reconciler_common.go:299] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/db6ae99f-f8a7-45ba-8336-fcd802d3eece-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.028788 5109 reconciler_common.go:299] "Volume detached for volume \"kube-api-access-kg4np\" (UniqueName: \"kubernetes.io/projected/db6ae99f-f8a7-45ba-8336-fcd802d3eece-kube-api-access-kg4np\") on node \"crc\" DevicePath \"\"" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.034603 5109 scope.go:117] "RemoveContainer" containerID="353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.098109 5109 scope.go:117] "RemoveContainer" containerID="e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb" Dec 11 17:16:44 crc kubenswrapper[5109]: E1211 17:16:44.098588 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb\": container with ID starting with e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb not found: ID does not exist" containerID="e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.098635 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb"} err="failed to get container status \"e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb\": rpc error: code = NotFound desc = could not find container \"e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb\": container with ID starting with e1075ab23b2172e6e52f6d6c86ff54d8677fcb1af658f8defbd8a34217ed77bb not found: ID does not exist" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.098662 5109 scope.go:117] "RemoveContainer" containerID="353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b" Dec 11 17:16:44 crc kubenswrapper[5109]: E1211 17:16:44.098982 5109 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b\": container with ID starting with 353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b not found: ID does not exist" containerID="353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.099011 5109 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b"} err="failed to get container status \"353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b\": rpc error: code = NotFound desc = could not find container \"353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b\": container with ID starting with 353ff6e81f9cd987c1b01906040e8ca479cea3945c54d094b049307b88e43c4b not found: ID does not exist" Dec 11 17:16:44 crc kubenswrapper[5109]: I1211 17:16:44.915276 5109 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db6ae99f-f8a7-45ba-8336-fcd802d3eece" path="/var/lib/kubelet/pods/db6ae99f-f8a7-45ba-8336-fcd802d3eece/volumes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116576245024461 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116576245017376 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116572731016515 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116572731015465 5ustar corecore